id
stringlengths
2
8
url
stringlengths
31
381
title
stringlengths
1
211
text
stringlengths
1
325k
12
https://en.wikipedia.org/wiki/Anarchism
Anarchism
Anarchism is a political philosophy and movement that is skeptical of all justifications for authority and seeks to abolish the institutions it claims maintain unnecessary coercion and hierarchy, typically including, though not necessarily limited to, governments, nation states, and capitalism. Anarchism advocates for the replacement of the state with stateless societies or other forms of free associations. As a historically left-wing movement, this reading of anarchism is placed on the farthest left of the political spectrum, it is usually described as the libertarian wing of the socialist movement (libertarian socialism). Humans have lived in societies without formal hierarchies long before the establishment of states, realms, or empires. With the rise of organised hierarchical bodies, scepticism toward authority also rose. Although traces of anarchist ideas are found all throughout history, modern anarchism emerged from the Enlightenment. During the latter half of the 19th and the first decades of the 20th century, the anarchist movement flourished in most parts of the world and had a significant role in workers' struggles for emancipation. Various anarchist schools of thought formed during this period. Anarchists have taken part in several revolutions, most notably in the Paris Commune, the Russian Civil War and the Spanish Civil War, whose end marked the end of the classical era of anarchism. In the last decades of the 20th and into the 21st century, the anarchist movement has been resurgent once more, growing in popularity and influence within anti-capitalist, anti-war and anti-globalisation movements. Anarchists employ a diversity of tactics in order to meet their ideal ends which can be broadly separated into revolutionary and evolutionary tactics; there is significant overlap between the two. Revolutionary tactics aim to bring down authority and state, having taken a violent turn in the past, while evolutionary tactics aim to prefigure what an anarchist society would be like. Anarchist thought, criticism, and praxis have played a part in diverse areas of human society. To achieve their ideal goals, anarchists use a variety of strategies, which may be generally divided into revolutionary and evolutionary strategies. Evolutionary methods try to simulate what an anarchist society might be like, but revolutionary tactics, which have historically taken a violent turn, aim to overthrow authority and the state. A variety of facets of human civilization have been influenced by anarchist theory, critique, and practice. Etymology, terminology, and definition The etymological origin of anarchism is from the Ancient Greek anarkhia, meaning "without a ruler", composed of the prefix an- ("without") and the word arkhos ("leader" or "ruler"). The suffix -ism denotes the ideological current that favours anarchy. Anarchism appears in English from 1642 as anarchisme and anarchy from 1539; early English usages emphasised a sense of disorder. Various factions within the French Revolution labelled their opponents as anarchists, although few such accused shared many views with later anarchists. Many revolutionaries of the 19th century such as William Godwin (1756–1836) and Wilhelm Weitling (1808–1871) would contribute to the anarchist doctrines of the next generation but did not use anarchist or anarchism in describing themselves or their beliefs. The first political philosopher to call himself an anarchist () was Pierre-Joseph Proudhon (1809–1865), marking the formal birth of anarchism in the mid-19th century. Since the 1890s and beginning in France, libertarianism has often been used as a synonym for anarchism and its use as a synonym is still common outside the United States. Some usages of libertarianism refer to individualistic free-market philosophy only, and free-market anarchism in particular is termed libertarian anarchism. While the term libertarian has been largely synonymous with anarchism, its meaning has more recently been diluted by wider adoption from ideologically disparate groups, including both the New Left and libertarian Marxists, who do not associate themselves with authoritarian socialists or a vanguard party, and extreme cultural liberals, who are primarily concerned with civil liberties. Additionally, some anarchists use libertarian socialist to avoid anarchism's negative connotations and emphasise its connections with socialism. Anarchism is broadly used to describe the anti-authoritarian wing of the socialist movement. Anarchism is contrasted to socialist forms which are state-oriented or from above. Scholars of anarchism generally highlight anarchism's socialist credentials and criticise attempts at creating dichotomies between the two. Some scholars describe anarchism as having many influences from liberalism, and being both liberal and socialist but more so, while most scholars reject anarcho-capitalism as a misunderstanding of anarchist principles. While opposition to the state is central to anarchist thought, defining anarchism is not an easy task for scholars, as there is a lot of discussion among scholars and anarchists on the matter, and various currents perceive anarchism slightly differently. Major definitional elements include the will for a non-coercive society, the rejection of the state apparatus, the belief that human nature allows humans to exist in or progress toward such a non-coercive society, and a suggestion on how to act to pursue the ideal of anarchy. History Pre-modern era Before the creation of towns and cities, established authority did not exist. It was after the institution of authority that anarchistic ideas were espoused as a reaction. The most notable precursors to anarchism in the ancient world were in China and Greece. In China, philosophical anarchism (the discussion on the legitimacy of the state) was delineated by Taoist philosophers Zhuang Zhou and Laozi. Alongside Stoicism, Taoism has been said to have had "significant anticipations" of anarchism. Anarchic attitudes were also articulated by tragedians and philosophers in Greece. Aeschylus and Sophocles used the myth of Antigone to illustrate the conflict between laws imposed by the state and personal autonomy. Socrates questioned Athenian authorities constantly and insisted on the right of individual freedom of conscience. Cynics dismissed human law (nomos) and associated authorities while trying to live according to nature (physis). Stoics were supportive of a society based on unofficial and friendly relations among its citizens without the presence of a state. In medieval Europe, there was no anarchistic activity except some ascetic religious movements. These, and other Muslim movements, later gave birth to religious anarchism. In the Sasanian Empire, Mazdak called for an egalitarian society and the abolition of monarchy, only to be soon executed by Emperor Kavad I. In Basra, religious sects preached against the state. In Europe, various sects developed anti-state and libertarian tendencies. Renewed interest in antiquity during the Renaissance and in private judgment during the Reformation restored elements of anti-authoritarian secularism, particularly in France. Enlightenment challenges to intellectual authority (secular and religious) and the revolutions of the 1790s and 1848 all spurred the ideological development of what became the era of classical anarchism. Modern era During the French Revolution, partisan groups such as the Enragés and the saw a turning point in the fermentation of anti-state and federalist sentiments. The first anarchist currents developed throughout the 18th century as William Godwin espoused philosophical anarchism in England, morally delegitimising the state, Max Stirner's thinking paved the way to individualism and Pierre-Joseph Proudhon's theory of mutualism found fertile soil in France. By the late 1870s, various anarchist schools of thought had become well-defined and a wave of then unprecedented globalisation occurred from 1880 to 1914. This era of classical anarchism lasted until the end of the Spanish Civil War and is considered the golden age of anarchism. Drawing from mutualism, Mikhail Bakunin founded collectivist anarchism and entered the International Workingmen's Association, a class worker union later known as the First International that formed in 1864 to unite diverse revolutionary currents. The International became a significant political force, with Karl Marx being a leading figure and a member of its General Council. Bakunin's faction (the Jura Federation) and Proudhon's followers (the mutualists) opposed state socialism, advocating political abstentionism and small property holdings. After bitter disputes, the Bakuninists were expelled from the International by the Marxists at the 1872 Hague Congress. Anarchists were treated similarly in the Second International, being ultimately expelled in 1896. Bakunin famously predicted that if revolutionaries gained power by Marx's terms, they would end up the new tyrants of workers. In response to their expulsion from the First International, anarchists formed the St. Imier International. Under the influence of Peter Kropotkin, a Russian philosopher and scientist, anarcho-communism overlapped with collectivism. Anarcho-communists, who drew inspiration from the 1871 Paris Commune, advocated for free federation and for the distribution of goods according to one's needs. By the turn of the 20th century, anarchism had spread all over the world. It was a notable feature of the international syndicalist movement. In China, small groups of students imported the humanistic pro-science version of anarcho-communism. Tokyo was a hotspot for rebellious youth from East Asian countries, who moved to the Japanese capital to study. In Latin America, Argentina was a stronghold for anarcho-syndicalism, where it became the most prominent left-wing ideology. During this time, a minority of anarchists adopted tactics of revolutionary political violence, known as propaganda of the deed. The dismemberment of the French socialist movement into many groups and the execution and exile of many Communards to penal colonies following the suppression of the Paris Commune favoured individualist political expression and acts. Even though many anarchists distanced themselves from these terrorist acts, infamy came upon the movement and attempts were made to prevent anarchists immigrating to the US, including the Immigration Act of 1903, also called the Anarchist Exclusion Act. Illegalism was another strategy which some anarchists adopted during this period. Despite concerns, anarchists enthusiastically participated in the Russian Revolution in opposition to the White movement, especially in the Makhnovshchina; however, they met harsh suppression after the Bolshevik government had stabilised, including during the Kronstadt rebellion. Several anarchists from Petrograd and Moscow fled to Ukraine, before the Bolsheviks crushed the anarchist movement there too. With the anarchists being repressed in Russia, two new antithetical currents emerged, namely platformism and synthesis anarchism. The former sought to create a coherent group that would push for revolution while the latter were against anything that would resemble a political party. Seeing the victories of the Bolsheviks in the October Revolution and the resulting Russian Civil War, many workers and activists turned to communist parties which grew at the expense of anarchism and other socialist movements. In France and the United States, members of major syndicalist movements such as the General Confederation of Labour and the Industrial Workers of the World left their organisations and joined the Communist International. In the Spanish Civil War of 1936–39, anarchists and syndicalists (CNT and FAI) once again allied themselves with various currents of leftists. A long tradition of Spanish anarchism led to anarchists playing a pivotal role in the war, and particularly in the Spanish Revolution of 1936. In response to the army rebellion, an anarchist-inspired movement of peasants and workers, supported by armed militias, took control of Barcelona and of large areas of rural Spain, where they collectivised the land. The Soviet Union provided some limited assistance at the beginning of the war, but the result was a bitter fight between communists and other leftists in a series of events known as the May Days, as Joseph Stalin asserted Soviet control of the Republican government, ending in another defeat of anarchists at the hands of the communists. Post-WWII By the end of World War II, the anarchist movement had been severely weakened. The 1960s witnessed a revival of anarchism, likely caused by a perceived failure of Marxism–Leninism and tensions built by the Cold War. During this time, anarchism found a presence in other movements critical towards both capitalism and the state such as the anti-nuclear, environmental, and peace movements, the counterculture of the 1960s, and the New Left. It also saw a transition from its previous revolutionary nature to provocative anti-capitalist reformism. Anarchism became associated with punk subculture as exemplified by bands such as Crass and the Sex Pistols. The established feminist tendencies of anarcha-feminism returned with vigour during the second wave of feminism. Black anarchism began to take form at this time and influenced anarchism's move from a Eurocentric demographic. This coincided with its failure to gain traction in Northern Europe and its unprecedented height in Latin America. Around the turn of the 21st century, anarchism grew in popularity and influence within anti-capitalist, anti-war and anti-globalisation movements. Anarchists became known for their involvement in protests against the World Trade Organization (WTO), the Group of Eight and the World Economic Forum. During the protests, ad hoc leaderless anonymous cadres known as black blocs engaged in rioting, property destruction and violent confrontations with the police. Other organisational tactics pioneered at this time include affinity groups, security culture and the use of decentralised technologies such as the Internet. A significant event of this period was the confrontations at the 1999 Seattle WTO conference. Anarchist ideas have been influential in the development of the Zapatistas in Mexico and the Democratic Federation of Northern Syria, more commonly known as Rojava, a de facto autonomous region in northern Syria. While having revolutionary aspirations, many forms of anarchism are not confrontational nowadays. Instead, they are trying to build an alternative way of social organization, based on mutual interdependence and voluntary cooperation. Scholar Carissa Honeywell takes the example of Food not Bombs group of collectives, to highlight some features of how anarchist groups work: direct action, working together and in solidarity with those left behind. While doing so, they inform about the rising rates of world hunger suggest a policy to tackle hunger, ranging from de-funding the arms industry to addressing Monsanto seed-saving policies and patents, helping farmers and commodification of food and housing. Honeywell also emphasizes that contemporary anarchists are interested in the flourishing not only of humans, but non-humans and the environment as well. Honeywell argues that escalation of problems such as continuous wars and world poverty show that the current framework not only cannot solve those pressing problems for humanity, but are causal factors as well, resulting in the rejection of representative democracy and the state as a whole. Branches Anarchist schools of thought have been generally grouped into two main historical traditions, social anarchism and individualist anarchism, owing to their different origins, values and evolution. The individualist current emphasises negative liberty in opposing restraints upon the free individual, while the social current emphasises positive liberty in aiming to achieve the free potential of society through equality and social ownership. In a chronological sense, anarchism can be segmented by the classical currents of the late 19th century and the post-classical currents (anarcha-feminism, green anarchism, and post-anarchism) developed thereafter. Beyond the specific factions of anarchist movements which constitute political anarchism lies philosophical anarchism which holds that the state lacks moral legitimacy, without necessarily accepting the imperative of revolution to eliminate it. A component especially of individualist anarchism, philosophical anarchism may tolerate the existence of a minimal state but claims that citizens have no moral obligation to obey government when it conflicts with individual autonomy. Anarchism pays significant attention to moral arguments since ethics have a central role in anarchist philosophy. Anarchism's emphasis on anti-capitalism, egalitarianism, and for the extension of community and individuality sets it apart from anarcho-capitalism and other types of economic libertarianism. Anarchism is usually placed on the far-left of the political spectrum. Much of its economics and legal philosophy reflect anti-authoritarian, anti-statist, libertarian, and radical interpretations of left-wing and socialist politics such as collectivism, communism, individualism, mutualism, and syndicalism, among other libertarian socialist economic theories. As anarchism does not offer a fixed body of doctrine from a single particular worldview, many anarchist types and traditions exist and varieties of anarchy diverge widely. One reaction against sectarianism within the anarchist milieu was anarchism without adjectives, a call for toleration and unity among anarchists first adopted by Fernando Tarrida del Mármol in 1889 in response to the bitter debates of anarchist theory at the time. Belief in political nihilism has been espoused by anarchists. Despite separation, the various anarchist schools of thought are not seen as distinct entities but rather as tendencies that intermingle and are connected through a set of uniform principles such as individual and local autonomy, mutual aid, network organisation, communal democracy, justified authority and decentralisation. Classical anarchism Inceptive currents among classical anarchist currents were mutualism and individualism. They were followed by the major currents of social anarchism (collectivist, communist and syndicalist). They differ on organisational and economic aspects of their ideal society. Mutualism is an 18th-century economic theory that was developed into anarchist theory by Pierre-Joseph Proudhon. Its aims include reciprocity, free association, voluntary contract, federation and monetary reform of both credit and currency that would be regulated by a bank of the people. Mutualism has been retrospectively characterised as ideologically situated between individualist and collectivist forms of anarchism. In What Is Property? (1840), Proudhon first characterised his goal as a "third form of society, the synthesis of communism and property." Collectivist anarchism is a revolutionary socialist form of anarchism commonly associated with Mikhail Bakunin. Collectivist anarchists advocate collective ownership of the means of production which is theorised to be achieved through violent revolution and that workers be paid according to time worked, rather than goods being distributed according to need as in communism. Collectivist anarchism arose alongside Marxism but rejected the dictatorship of the proletariat despite the stated Marxist goal of a collectivist stateless society. Anarcho-communism is a theory of anarchism that advocates a communist society with common ownership of the means of production, held by a federal network of voluntary associations, with production and consumption based on the guiding principle "From each according to his ability, to each according to his need." Anarcho-communism developed from radical socialist currents after the French Revolution but was first formulated as such in the Italian section of the First International. It was later expanded upon in the theoretical work of Peter Kropotkin, whose specific style would go onto become the dominating view of anarchists by the late 19th century. Anarcho-syndicalism is a branch of anarchism that views labour syndicates as a potential force for revolutionary social change, replacing capitalism and the state with a new society democratically self-managed by workers. The basic principles of anarcho-syndicalism are direct action, workers' solidarity and workers' self-management. Individualist anarchism is a set of several traditions of thought within the anarchist movement that emphasise the individual and their will over any kinds of external determinants. Early influences on individualist forms of anarchism include William Godwin, Max Stirner, and Henry David Thoreau. Through many countries, individualist anarchism attracted a small yet diverse following of Bohemian artists and intellectuals as well as young anarchist outlaws in what became known as illegalism and individual reclamation. Contemporary anarchism Anarchist principles undergird contemporary radical social movements of the left. Interest in the anarchist movement developed alongside momentum in the anti-globalisation movement, whose leading activist networks were anarchist in orientation. As the movement shaped 21st century radicalism, wider embrace of anarchist principles signaled a revival of interest. Anarchism has continued to generate many philosophies and movements, at times eclectic, drawing upon various sources and combining disparate concepts to create new philosophical approaches. The anti-capitalist tradition of classical anarchism has remained prominent within contemporary currents. Contemporary news coverage which emphasizes black bloc demonstrations has reinforced anarchism's historical association with chaos and violence. Its publicity has also led more scholars in fields such as anthropology and history to engage with the anarchist movement, although contemporary anarchism favours actions over academic theory. Various anarchist groups, tendencies, and schools of thought exist today, making it difficult to describe the contemporary anarchist movement. While theorists and activists have established "relatively stable constellations of anarchist principles", there is no consensus on which principles are core and commentators describe multiple anarchisms, rather than a singular anarchism, in which common principles are shared between schools of anarchism while each group prioritizes those principles differently. Gender equality can be a common principle, although it ranks as a higher priority to anarcha-feminists than anarcho-communists. Anarchists are generally committed against coercive authority in all forms, namely "all centralized and hierarchical forms of government (e.g., monarchy, representative democracy, state socialism, etc.), economic class systems (e.g., capitalism, Bolshevism, feudalism, slavery, etc.), autocratic religions (e.g., fundamentalist Islam, Roman Catholicism, etc.), patriarchy, heterosexism, white supremacy, and imperialism." Anarchist schools disagree on the methods by which these forms should be opposed. The principle of equal liberty is closer to anarchist political ethics in that it transcends both the liberal and socialist traditions. This entails that liberty and equality cannot be implemented within the state, resulting in the questioning of all forms of domination and hierarchy. Tactics Anarchists' tactics take various forms but in general serve two major goals, namely, to first oppose the Establishment and secondly to promote anarchist ethics and reflect an anarchist vision of society, illustrating the unity of means and ends. A broad categorisation can be made between aims to destroy oppressive states and institutions by revolutionary means on one hand and aims to change society through evolutionary means on the other. Evolutionary tactics embrace nonviolence, reject violence and take a gradual approach to anarchist aims, although there is significant overlap between the two. Anarchist tactics have shifted during the course of the last century. Anarchists during the early 20th century focused more on strikes and militancy while contemporary anarchists use a broader array of approaches. Classical era During the classical era, anarchists had a militant tendency. Not only did they confront state armed forces, as in Spain and Ukraine, but some of them also employed terrorism as propaganda of the deed. Assassination attempts were carried out against heads of state, some of which were successful. Anarchists also took part in revolutions. Many anarchists, especially the Galleanists, believed that these attempts would be the impetus for a revolution against capitalism and the state. Many of these attacks were done by individual assailants and the majority took place in the late 1870s, the early 1880s and the 1890s, with some still occurring in the early 1900s. Their decrease in prevalence was the result of further judicial power and targeting and cataloging by state institutions. Anarchist perspectives towards violence have always been controversial. Anarcho-pacifists advocate for non-violence means to achieve their stateless, nonviolent ends. Other anarchist groups advocate direct action, a tactic which can include acts of sabotage or terrorism. This attitude was quite prominent a century ago when seeing the state as a tyrant and some anarchists believing that they had every right to oppose its oppression by any means possible. Emma Goldman and Errico Malatesta, who were proponents of limited use of violence, stated that violence is merely a reaction to state violence as a necessary evil. Anarchists took an active role in strike actions, although they tended to be antipathetic to formal syndicalism, seeing it as reformist. They saw it as a part of the movement which sought to overthrow the state and capitalism. Anarchists also reinforced their propaganda within the arts, some of whom practiced naturism and nudism. Those anarchists also built communities which were based on friendship and were involved in the news media. Revolutionary In the current era, Italian anarchist Alfredo Bonanno, a proponent of insurrectionary anarchism, has reinstated the debate on violence by rejecting the nonviolence tactic adopted since the late 19th century by Kropotkin and other prominent anarchists afterwards. Both Bonanno and the French group The Invisible Committee advocate for small, informal affiliation groups, where each member is responsible for their own actions but works together to bring down oppression utilizing sabotage and other violent means against state, capitalism, and other enemies. Members of The Invisible Committee were arrested in 2008 on various charges, terrorism included. Overall, contemporary anarchists are much less violent and militant than their ideological ancestors. They mostly engage in confronting the police during demonstrations and riots, especially in countries such as Canada, Greece, and Mexico. Militant black bloc protest groups are known for clashing with the police; however, anarchists not only clash with state operators, they also engage in the struggle against fascists and racists, taking anti-fascist action and mobilizing to prevent hate rallies from happening. Evolutionary Anarchists commonly employ direct action. This can take the form of disrupting and protesting against unjust hierarchy, or the form of self-managing their lives through the creation of counter-institutions such as communes and non-hierarchical collectives. Decision-making is often handled in an anti-authoritarian way, with everyone having equal say in each decision, an approach known as horizontalism. Contemporary-era anarchists have been engaging with various grassroots movements that are more or less based on horizontalism, although not explicitly anarchist, respecting personal autonomy and participating in mass activism such as strikes and demonstrations. In contrast with the big-A anarchism of the classical era, the newly coined term small-a anarchism signals their tendency not to base their thoughts and actions on classical-era anarchism or to refer to classical anarchists such as Peter Kropotkin and Pierre-Joseph Proudhon to justify their opinions. Those anarchists would rather base their thought and praxis on their own experience which they will later theorize. The decision-making process of small anarchist affinity groups plays a significant tactical role. Anarchists have employed various methods in order to build a rough consensus among members of their group without the need of a leader or a leading group. One way is for an individual from the group to play the role of facilitator to help achieve a consensus without taking part in the discussion themselves or promoting a specific point. Minorities usually accept rough consensus, except when they feel the proposal contradicts anarchist ethics, goals and values. Anarchists usually form small groups (5–20 individuals) to enhance autonomy and friendships among their members. These kinds of groups more often than not interconnect with each other, forming larger networks. Anarchists still support and participate in strikes, especially wildcat strikes as these are leaderless strikes not organised centrally by a syndicate. As in the past, newspapers and journals are used, and anarchists have gone online in the World Wide Web to spread their message. Anarchists have found it easier to create websites because of distributional and other difficulties, hosting electronic libraries and other portals. Anarchists were also involved in developing various software that are available for free. The way these hacktivists work to develop and distribute resembles the anarchist ideals, especially when it comes to preserving users' privacy from state surveillance. Anarchists organize themselves to squat and reclaim public spaces. During important events such as protests and when spaces are being occupied, they are often called Temporary Autonomous Zones (TAZ), spaces where art, poetry, and surrealism are blended to display the anarchist ideal. As seen by anarchists, squatting is a way to regain urban space from the capitalist market, serving pragmatical needs and also being an exemplary direct action. Acquiring space enables anarchists to experiment with their ideas and build social bonds. Adding up these tactics while having in mind that not all anarchists share the same attitudes towards them, along with various forms of protesting at highly symbolic events, make up a carnivalesque atmosphere that is part of contemporary anarchist vividity. Key issues As anarchism is a philosophy that embodies many diverse attitudes, tendencies, and schools of thought; disagreement over questions of values, ideology, and tactics is common. Its diversity has led to widely different uses of identical terms among different anarchist traditions which has created a number of definitional concerns in anarchist theory. The compatibility of capitalism, nationalism, and religion with anarchism is widely disputed, and anarchism enjoys complex relationships with ideologies such as communism, collectivism, Marxism, and trade unionism. Anarchists may be motivated by humanism, divine authority, enlightened self-interest, veganism, or any number of alternative ethical doctrines. Phenomena such as civilisation, technology (e.g. within anarcho-primitivism), and the democratic process may be sharply criticised within some anarchist tendencies and simultaneously lauded in others. The state Objection to the state and its institutions is a sine qua non of anarchism. Anarchists consider the state as a tool of domination and believe it to be illegitimate regardless of its political tendencies. Instead of people being able to control the aspects of their life, major decisions are taken by a small elite. Authority ultimately rests solely on power, regardless of whether that power is open or transparent, as it still has the ability to coerce people. Another anarchist argument against states is that the people constituting a government, even the most altruistic among officials, will unavoidably seek to gain more power, leading to corruption. Anarchists consider the idea that the state is the collective will of the people to be an unachievable fiction due to the fact that the ruling class is distinct from the rest of society. Specific anarchist attitudes towards the state vary. Robert Paul Wolff believed that the tension between authority and autonomy would mean the state could never be legitimate. Bakunin saw the state as meaning "coercion, domination by means of coercion, camouflaged if possible but unceremonious and overt if need be." A. John Simmons and Leslie Green, who leaned toward philosophical anarchism, believed that the state could be legitimate if it is governed by consensus, although they saw this as highly unlikely. Beliefs on how to abolish the state also differ. Gender, sexuality, and free love As gender and sexuality carry along them dynamics of hierarchy, many anarchists address, analyse, and oppose the suppression of one's autonomy imposed by gender roles. Sexuality was not often discussed by classical anarchists but the few that did felt that an anarchist society would lead to sexuality naturally developing. Sexual violence was a concern for anarchists such as Benjamin Tucker, who opposed age of consent laws, believing they would benefit predatory men. A historical current that arose and flourished during 1890 and 1920 within anarchism was free love. In contemporary anarchism, this current survives as a tendency to support polyamory, relationship anarchy, and queer anarchism. Free love advocates were against marriage, which they saw as a way of men imposing authority over women, largely because marriage law greatly favoured the power of men. The notion of free love was much broader and included a critique of the established order that limited women's sexual freedom and pleasure. Those free love movements contributed to the establishment of communal houses, where large groups of travelers, anarchists and other activists slept in beds together. Free love had roots both in Europe and the United States; however, some anarchists struggled with the jealousy that arose from free love. Anarchist feminists were advocates of free love, against marriage, and pro-choice (utilising a contemporary term), and had a similar agenda. Anarchist and non-anarchist feminists differed on suffrage but were supportive of one another. During the second half of the 20th century, anarchism intermingled with the second wave of feminism, radicalising some currents of the feminist movement and being influenced as well. By the latest decades of the 20th century, anarchists and feminists were advocating for the rights and autonomy of women, gays, queers and other marginalised groups, with some feminist thinkers suggesting a fusion of the two currents. With the third wave of feminism, sexual identity and compulsory heterosexuality became a subject of study for anarchists, yielding a post-structuralist critique of sexual normality. Some anarchists distanced themselves from this line of thinking, suggesting that it leaned towards an individualism that was dropping the cause of social liberation. Education The interest of anarchists in education stretches back to the first emergence of classical anarchism. Anarchists consider proper education, one which sets the foundations of the future autonomy of the individual and the society, to be an act of mutual aid. Anarchist writers such as William Godwin (Political Justice) and Max Stirner ("The False Principle of Our Education") attacked both state education and private education as another means by which the ruling class replicate their privileges. In 1901, Catalan anarchist and free thinker Francisco Ferrer established the Escuela Moderna in Barcelona as an opposition to the established education system which was dictated largely by the Catholic Church. Ferrer's approach was secular, rejecting both state and church involvement in the educational process whilst giving pupils large amounts of autonomy in planning their work and attendance. Ferrer aimed to educate the working class and explicitly sought to foster class consciousness among students. The school closed after constant harassment by the state and Ferrer was later arrested. Nonetheless, his ideas formed the inspiration for a series of modern schools around the world. Christian anarchist Leo Tolstoy, who published the essay Education and Culture, also established a similar school with its founding principle being that "for education to be effective it had to be free." In a similar token, A. S. Neill founded what became the Summerhill School in 1921, also declaring being free from coercion. Anarchist education is based largely on the idea that a child's right to develop freely and without manipulation ought to be respected and that rationality would lead children to morally good conclusions; however, there has been little consensus among anarchist figures as to what constitutes manipulation. Ferrer believed that moral indoctrination was necessary and explicitly taught pupils that equality, liberty and social justice were not possible under capitalism, along with other critiques of government and nationalism. Late 20th century and contemporary anarchist writers (Paul Goodman, Herbert Read, and Colin Ward) intensified and expanded the anarchist critique of state education, largely focusing on the need for a system that focuses on children's creativity rather than on their ability to attain a career or participate in consumerism as part of a consumer society. Contemporary anarchists such as Ward claim that state education serves to perpetuate socioeconomic inequality. While few anarchist education institutions have survived to the modern-day, major tenets of anarchist schools, among them respect for child autonomy and relying on reasoning rather than indoctrination as a teaching method, have spread among mainstream educational institutions. Judith Suissa names three schools as explicitly anarchists' schools, namely the Free Skool Santa Cruz in the United States which is part of a wider American-Canadian network of schools, the Self-Managed Learning College in Brighton, England, and the Paideia School in Spain. The arts The connection between anarchism and art was quite profound during the classical era of anarchism, especially among artistic currents that were developing during that era such as futurists, surrealists and others. In literature, anarchism was mostly associated with the New Apocalyptics and the neo-romanticism movement. In music, anarchism has been associated with music scenes such as punk. Anarchists such as Leo Tolstoy and Herbert Read stated that the border between the artist and the non-artist, what separates art from a daily act, is a construct produced by the alienation caused by capitalism and it prevents humans from living a joyful life. Other anarchists advocated for or used art as a means to achieve anarchist ends. In his book Breaking the Spell: A History of Anarchist Filmmakers, Videotape Guerrillas, and Digital Ninjas, Chris Robé claims that "anarchist-inflected practices have increasingly structured movement-based video activism." Throughout the 20th century, many prominent anarchists (Peter Kropotkin, Emma Goldman, Gustav Landauer and Camillo Berneri) and publications such as Anarchy wrote about matters pertaining to the arts. Three overlapping properties made art useful to anarchists. It could depict a critique of existing society and hierarchies, serve as a prefigurative tool to reflect the anarchist ideal society and even turn into a means of direct action such as in protests. As it appeals to both emotion and reason, art could appeal to the whole human and have a powerful effect. The 19th-century neo-impressionist movement had an ecological aesthetic and offered an example of an anarchist perception of the road towards socialism. In Les chataigniers a Osny by anarchist painter Camille Pissarro, the blending of aesthetic and social harmony is prefiguring an ideal anarchistic agrarian community. Criticism The most common critique of anarchism is the assertion that humans cannot self-govern and so a state is necessary for human survival. Philosopher Bertrand Russell supported this critique, stating that "[p]eace and war, tariffs, regulations of sanitary conditions and the sale of noxious drugs, the preservation of a just system of distribution: these, among others, are functions which could hardly be performed in a community in which there was no central government." Another common criticism of anarchism is that it fits a world of isolation in which only the small enough entities can be self-governing; a response would be that major anarchist thinkers advocated anarchist federalism. Philosophy lecturer Andrew G. Fiala composed a list of common arguments against anarchism which includes critiques such as that anarchism is innately related to violence and destruction, not only in the pragmatic world, such as at protests, but in the world of ethics as well. Secondly, anarchism is evaluated as unfeasible or utopian since the state cannot be defeated practically. This line of arguments most often calls for political action within the system to reform it. The third argument is that anarchism is self-contradictory as a ruling theory that has no ruling theory. Anarchism also calls for collective action whilst endorsing the autonomy of the individual, hence no collective action can be taken. Lastly, Fiala mentions a critique towards philosophical anarchism of being ineffective (all talk and thoughts) and in the meantime capitalism and bourgeois class remains strong. Philosophical anarchism has met the criticism of members of academia following the release of pro-anarchist books such as A. John Simmons' Moral Principles and Political Obligations. Law professor William A. Edmundson authored an essay to argue against three major philosophical anarchist principles which he finds fallacious. Edmundson says that while the individual does not owe the state a duty of obedience, this does not imply that anarchism is the inevitable conclusion and the state is still morally legitimate. In The Problem of Political Authority, Michael Huemer defends philosophical anarchism, claiming that "political authority is a moral illusion." One of the earliest criticisms is that anarchism defies and fails to understand the biological inclination to authority. Joseph Raz states that the acceptance of authority implies the belief that following their instructions will afford more success. Raz believes that this argument is true in following both authorities' successful and mistaken instruction. Anarchists reject this criticism because challenging or disobeying authority does not entail the disappearance of its advantages by acknowledging authority such as doctors or lawyers as reliable, nor does it involve a complete surrender of independent judgment. Anarchist perception of human nature, rejection of the state, and commitment to social revolution has been criticised by academics as naive, overly simplistic, and unrealistic, respectively. Classical anarchism has been criticised for relying too heavily on the belief that the abolition of the state will lead to human cooperation prospering. Friedrich Engels, considered to be one of the principal founders of Marxism, criticised anarchism's anti-authoritarianism as inherently counter-revolutionary because in his view a revolution is by itself authoritarian. Academic John Molyneux writes in his book Anarchism: A Marxist Criticism that "anarchism cannot win", believing that it lacks the ability to properly implement its ideas. The Marxist criticism of anarchism is that it has a utopian character because all individuals should have anarchist views and values. According to the Marxist view, that a social idea would follow directly from this human ideal and out of the free will of every individual formed its essence. Marxists state that this contradiction was responsible for their inability to act. In the anarchist vision, the conflict between liberty and equality was resolved through coexistence and intertwining. See also Anarchism by country Governance without government List of anarchist political ideologies List of books about anarchism References Explanatory notes Citations General and cited sources Primary sources Secondary sources Tertiary sources Further reading Criticism of philosophical anarchism. A defence of philosophical anarchism, stating that "both kinds of 'anarchism' [i.e. philosophical and political anarchism] are philosophical and political claims." (p. 137) Anarchistic popular fiction novel. An argument for philosophical anarchism. External links Anarchy Archives – an online research center on the history and theory of anarchism. Anti-capitalism Anti-fascism Economic ideologies Far-left politics Left-wing politics Libertarian socialism Libertarianism Political culture Political ideologies Political movements Social theories Socialism
709
https://en.wikipedia.org/wiki/Angolan%20Armed%20Forces
Angolan Armed Forces
The Angolan Armed Forces () or FAA is the military of Angola. The FAA consist of the Angolan Army (), the Angolan Navy () and the National Air Force of Angola (). Reported total manpower in 2021 was about 107,000. The FAA is headed by the Chief of the General Staff António Egídio de Sousa Santos since 2018, who reports to the minister of National Defense, currently João Ernesto dos Santos. History Roots The FAA succeeded to the previous People's Armed Forces for the Liberation of Angola (FAPLA) following the abortive Bicesse Accord with the Armed Forces of the Liberation of Angola (FALA), armed wing of the National Union for the Total Independence of Angola (UNITA). As part of the peace agreement, troops from both armies were to be demilitarized and then integrated. Integration was never completed as UNITA and FALA went back to war in 1992. Later, consequences for FALA personnel in Luanda were harsh with FAPLA veterans persecuting their erstwhile opponents in certain areas and reports of vigilantism. Founding The Angolan Armed Forces were created on 9 October 1991. The institutionalization of the FAA was made in the Bicesse Accords, signed in 1991, between the Angolan Government and UNITA. The principles that would govern the FAA were defined in a joint proposal presented on September 24, 1991 and approved on 9 October. On 14 November 1991, Generals João Baptista de Matos and Abílio Kamalata Numa were appointed to the Superior Command of the Armed Forces. The ceremony took place at the Hotel Presidente Luanda, and was presided over by the then-minister França Vandúnem. Branches Army The Army (Exército) is the land component of the FAA. It is organized in six military regions (Cabinda, Luanda, North, Center, East and South), with an infantry division being based in each one. Distributed by the six military regions / infantry divisions, there are 25 motorized infantry brigades, one tank brigade and one engineering brigade. The Army also includes an artillery regiment, the Military Artillery School, the Army Military Academy, an anti-aircraft defense group, a composite land artillery group, a military police regiment, a logistical transportation regiment and a field artillery brigade. The Army further includes the Special Forces Brigade (including Commandos and Special Operations units), but this unit is under the direct command of the General Staff of the FAA. Air Force The National Air Force of Angola (FANA, Força Aérea Nacional de Angola) is the air component of the FAA. It is organized in six aviation regiments, each including several squadrons. To each of the regiments correspond an air base. Besides the aviation regiments, there is also a Pilot Training School. The Air Force's personnel total about 8,000; its equipment includes transport aircraft and six Russian-manufactured Sukhoi Su-27 fighter aircraft. In 2002, one was lost during the civil war with UNITA forces. In 1991, the Air Force/Air Defense Forces had 8,000 personnel and 90 combat-capable aircraft, including 22 fighters, 59 fighter ground attack aircraft and 16 attack helicopters. Navy The Angola Navy (MGA, Marinha de Guerra de Angola) is the naval component of the FAA. It is organized in two naval zones (North and South), with naval bases in Luanda, Lobito and Moçâmedes. It includes a Marines Brigade and a Marines School, based in Ambriz. The Navy numbers about 1,000 personnel and operates only a handful of small patrol craft and barges. The Navy has been neglected and ignored as a military arm mainly due to the guerrilla struggle against the Portuguese and the nature of the civil war. From the early 1990s to the present the Angolan Navy has shrunk from around 4,200 personnel to around 1,000, resulting in the loss of skills and expertise needed to maintain equipment. In order to protect Angola's 1 600 km long coastline, the Angolan Navy is undergoing modernisation but is still lacking in many ways. Portugal has been providing training through its Technical Military Cooperation (CTM) programme. The Navy is requesting procurement of a frigate, three corvettes, three offshore patrol vessel and additional fast patrol boats. Most of the vessels in the navy's inventory dates back from the 1980s or earlier, and many of its ships are inoperable due to age and lack of maintenance. However the navy acquired new boats from Spain and France in the 1990s. Germany has delivered several Fast Attack Craft for border protection in 2011. In September 2014 it was reported that the Angolan Navy would acquire seven Macaé-class patrol vessels from Brazil as part of a Technical Memorandum of Understanding (MoU) covering the production of the vessels as part of Angola's Naval Power Development Programme (Pronaval). The military of Angola aims to modernize its naval capability, presumably due to a rise in maritime piracy within the Gulf of Guinea which may have an adverse effect on the country's economy. The navy's current known inventory includes the following: Fast attack craft 4 Mandume class craft (Bazan Cormoran type, refurbished in 2009) Patrol Boats 3 18.3m long Patrulheiro patrol boats (refurbished in 2002) 5 ARESA PVC-170 2 Namacurra-class harbour patrol boats Fisheries Patrol Boats Ngola Kiluange and Nzinga Mbandi (delivered in September and October 2012 from Damen Shipyards)(Operated by Navy personnel under the Ministry of Agriculture, Rural Development and Fisheries) 28 metre FRV 2810 (Pensador) (Operated by Navy personnel under the Ministry of Agriculture, Rural Development and Fisheries) Landing craft LDM-400 - 1 or 3 (reportedly has serviceability issues) Coastal defense equipment (CRTOC) SS-C1 Sepal radar system The navy also has several aircraft for maritime patrol: Specialized units Special forces The FAA include several types of special forces, namely the Commandos, the Special Operations and the Marines. The Angolan special forces follow the general model of the analogous Portuguese special forces, receiving a similar training. The Commandos and the Special forces are part of the Special Forces Brigade (BRIFE, Brigada de Forças Especiais), based at Cabo Ledo, in the Bengo Province. The BRIFE includes two battalions of commandos, a battalion of special operations and sub-units of combat support and service support. The BRIFE also included the Special Actions Group (GAE, Grupo de Ações Especiais), which is presently inactive and that was dedicated to long range reconnaissance, covert and sabotage operations. In the Cabo Ledo base is also installed the Special Forces Training School (EFFE, Escola de Formação de Forças Especiais). Both the BRIFE and the EFFE are directly under the Directorate of Special Forces of the General Staff of the Armed Forces. The marines (fuzileiros navais) constitute the Marines Brigade of the Angolan Navy. The Marines Brigade is not permanently dependent of the Directorate of Special Forces, but can detach their units and elements to be put under the command of that body for the conduction of exercises or real operations. Since the disbandment of the Angolan Parachute Battalion in 2004, the FAA do not have a specialized paratrooper unit. However, elements of the commandos, special operations and marines are parachute qualified. Territorial troops The Directorate of People's Defense and Territorial Troops of the Defence Ministry or ODP was established in late 1975. It had 600,000 members, having personnel in virtually every village by 1979. It had both armed and unarmed units dispersed in villages throughout the country. The People's Vigilance Brigades () also serve a similar purpose. Training establishments Armed Forces Academy The Military Academy () is a military university public higher education establishment whose mission is to train officers of the Permanent Staff of the Army. It has been in operation since 21 August 2009 by presidential decree. Its headquarters are in Lobito. It trains in the following specialties: Infantry Tanks Land Artillery Anti-Air Defense Military Engineering Logistics Telecommunications Hidden Direction of Troops Military Administration Armament and Technique Chemical Defense Operational Military Intelligence Technical Repair and Maintenance Platoon of Auto and Armored Technique Navy Naval War Institute (INSG) Naval Academy Naval Specialist School Air Force Angolan Military Aviation School Pilot Basic Training School (Lobito) Institutions/other units Museum of the Armed Forces Military Hospitals The Military hospital of the FAA is the Main Military Hospital. It has the following lineage: 1961 – Evacuation Infirmary 1962 – Military Hospital of Luanda 1975 – Military Hospital 1976 – Central Military Hospital 1989 – Main Military Hospital It provides specialized medical assistance in accordance with the military health system; It also promotes post-graduate education and scientific research. Currently, the Main Military Hospital serves 39 special medical specialties. It is a headed by a Director General whose main supporting body is the Board of Directors. Supreme Military Court The Supreme Military Court is the highest organ of the hierarchy of military courts. The Presiding Judge, the Deputy Presiding Judge and the other Counselor Judges of the Supreme Military Court are appointed by the President of the Republic. The composition, organization, powers and functioning of the Supreme Military Court are established by law. Military Bands The FAA maintains Portuguese-style military bands in all three branches and in individual units. The primary band is the 100-member Music Band of the Presidential Security Household. The music band of the Army Command was created on 16 June 1994 and four years later, on 15 August 1998, the National Air Force created a music band within an artistic brigade. The navy has its own marching band, as well as a small musical group known as Banda 10 de Julho (10th July Band), based at the Luanda Naval Base. Foreign deployments The FAPLA's main counterinsurgency effort was directed against UNITA in the southeast, and its conventional capabilities were demonstrated principally in the undeclared South African Border War. The FAPLA first performed its external assistance mission with the dispatch of 1,000 to 1,500 troops to São Tomé and Príncipe in 1977 to bolster the socialist regime of President Manuel Pinto da Costa. During the next several years, Angolan forces conducted joint exercises with their counterparts and exchanged technical operational visits. The Angolan expeditionary force was reduced to about 500 in early 1985. The Angolan Armed Forces were controversially involved in training the armed forces of fellow Lusophone states Cape Verde and Guinea-Bissau. In the case of the latter, the 2012 Guinea-Bissau coup d'état was cited by the coup leaders as due to Angola's involvement in trying to "reform" the military in connivance with the civilian leadership. A small number of FAA personnel are stationed in the Democratic Republic of the Congo (Kinshasa). A presence during the unrest in Ivory Coast, 2010–2011, were not officially confirmed. However, the Frankfurter Allgemeine Zeitung, citing Jeune Afrique, said that among President Gbagbo's guards were 92 personnel of President Dos Santos's Presidential Guard Unit. Angola is basically interested in the participation of the FAA operations of the African Union and has formed special units for this purpose. In 2021, the Angolan Parliament approved integration of FAA into Southern African Development Community (SADC)'s mission for peace in Cabo Delgado, Mozambique. Angola sent a team of 20 officers to participate. References Further reading Human Rights Watch, Angola Unravels: The Rise and Fall of the Lusaka Peace Process, October 1999 Utz Ebertz and Marie Müller, Legacy of a resource-fueled war: The role of generals in Angola's mining sector, BICC Focus, June 2013 Area Handbook for Angola, August 1967, Angola, A Country Study (1979 and 1991) Rocky Williams, "National defence reform and the African Union." SIPRI Yearbook 2004: 231–249. Weigert, Stephen L. Angola: a modern military history, 1961–2002. Palgrave Macmillan, 2011. Martin Rupiya et al., 'Angola', in Evolutions and Revolutions The Twenty-Seventh of May: An Historical Note on the Abortive 1977 "coup" in Angola David Birmingham, African Affairs, Vol. 77, No. 309 (Oct. 1978), pp. 554–564 Published by: Oxford University Press on behalf of The Royal African Society External links Official site of the Angolan Ministry of National Defence World Navies Brinkman, Inge "Language, Names, and War: The Case of Angola", African Studies Review Military of Angola Military history of Angola Angolan Civil War 1991 establishments in Angola Military units and formations established in 1991
900
https://en.wikipedia.org/wiki/Americium
Americium
Americium is a synthetic radioactive chemical element with the symbol Am and atomic number 95. It is a transuranic member of the actinide series, in the periodic table located under the lanthanide element europium, and thus by analogy was named after the Americas. Americium was first produced in 1944 by the group of Glenn T. Seaborg from Berkeley, California, at the Metallurgical Laboratory of the University of Chicago, as part of the Manhattan Project. Although it is the third element in the transuranic series, it was discovered fourth, after the heavier curium. The discovery was kept secret and only released to the public in November 1945. Most americium is produced by uranium or plutonium being bombarded with neutrons in nuclear reactors – one tonne of spent nuclear fuel contains about 100 grams of americium. It is widely used in commercial ionization chamber smoke detectors, as well as in neutron sources and industrial gauges. Several unusual applications, such as nuclear batteries or fuel for space ships with nuclear propulsion, have been proposed for the isotope 242mAm, but they are as yet hindered by the scarcity and high price of this nuclear isomer. Americium is a relatively soft radioactive metal with silvery appearance. Its most common isotopes are 241Am and 243Am. In chemical compounds, americium usually assumes the oxidation state +3, especially in solutions. Several other oxidation states are known, ranging from +2 to +7, and can be identified by their characteristic optical absorption spectra. The crystal lattices of solid americium and its compounds contain small intrinsic radiogenic defects, due to metamictization induced by self-irradiation with alpha particles, which accumulates with time; this can cause a drift of some material properties over time, more noticeable in older samples. History Although americium was likely produced in previous nuclear experiments, it was first intentionally synthesized, isolated and identified in late autumn 1944, at the University of California, Berkeley, by Glenn T. Seaborg, Leon O. Morgan, Ralph A. James, and Albert Ghiorso. They used a 60-inch cyclotron at the University of California, Berkeley. The element was chemically identified at the Metallurgical Laboratory (now Argonne National Laboratory) of the University of Chicago. Following the lighter neptunium, plutonium, and heavier curium, americium was the fourth transuranium element to be discovered. At the time, the periodic table had been restructured by Seaborg to its present layout, containing the actinide row below the lanthanide one. This led to americium being located right below its twin lanthanide element europium; it was thus by analogy named after the Americas: "The name americium (after the Americas) and the symbol Am are suggested for the element on the basis of its position as the sixth member of the actinide rare-earth series, analogous to europium, Eu, of the lanthanide series." The new element was isolated from its oxides in a complex, multi-step process. First plutonium-239 nitrate (239PuNO3) solution was coated on a platinum foil of about 0.5 cm2 area, the solution was evaporated and the residue was converted into plutonium dioxide (PuO2) by calcining. After cyclotron irradiation, the coating was dissolved with nitric acid, and then precipitated as the hydroxide using concentrated aqueous ammonia solution. The residue was dissolved in perchloric acid. Further separation was carried out by ion exchange, yielding a certain isotope of curium. The separation of curium and americium was so painstaking that those elements were initially called by the Berkeley group as pandemonium (from Greek for all demons or hell) and delirium (from Latin for madness). Initial experiments yielded four americium isotopes: 241Am, 242Am, 239Am and 238Am. Americium-241 was directly obtained from plutonium upon absorption of two neutrons. It decays by emission of a α-particle to 237Np; the half-life of this decay was first determined as years but then corrected to 432.2 years. The times are half-lives The second isotope 242Am was produced upon neutron bombardment of the already-created 241Am. Upon rapid β-decay, 242Am converts into the isotope of curium 242Cm (which had been discovered previously). The half-life of this decay was initially determined at 17 hours, which was close to the presently accepted value of 16.02 h. The discovery of americium and curium in 1944 was closely related to the Manhattan Project; the results were confidential and declassified only in 1945. Seaborg leaked the synthesis of the elements 95 and 96 on the U.S. radio show for children Quiz Kids five days before the official presentation at an American Chemical Society meeting on 11 November 1945, when one of the listeners asked whether any new transuranium element besides plutonium and neptunium had been discovered during the war. After the discovery of americium isotopes 241Am and 242Am, their production and compounds were patented listing only Seaborg as the inventor. The initial americium samples weighed a few micrograms; they were barely visible and were identified by their radioactivity. The first substantial amounts of metallic americium weighing 40–200 micrograms were not prepared until 1951 by reduction of americium(III) fluoride with barium metal in high vacuum at 1100 °C. Occurrence The longest-lived and most common isotopes of americium, 241Am and 243Am, have half-lives of 432.2 and 7,370 years, respectively. Therefore, any primordial americium (americium that was present on Earth during its formation) should have decayed by now. Trace amounts of americium probably occur naturally in uranium minerals as a result of nuclear reactions, though this has not been confirmed. Existing americium is concentrated in the areas used for the atmospheric nuclear weapons tests conducted between 1945 and 1980, as well as at the sites of nuclear incidents, such as the Chernobyl disaster. For example, the analysis of the debris at the testing site of the first U.S. hydrogen bomb, Ivy Mike, (1 November 1952, Enewetak Atoll), revealed high concentrations of various actinides including americium; but due to military secrecy, this result was not published until later, in 1956. Trinitite, the glassy residue left on the desert floor near Alamogordo, New Mexico, after the plutonium-based Trinity nuclear bomb test on 16 July 1945, contains traces of americium-241. Elevated levels of americium were also detected at the crash site of a US Boeing B-52 bomber aircraft, which carried four hydrogen bombs, in 1968 in Greenland. In other regions, the average radioactivity of surface soil due to residual americium is only about 0.01 picocuries per gram (0.37 mBq/g). Atmospheric americium compounds are poorly soluble in common solvents and mostly adhere to soil particles. Soil analysis revealed about 1,900 times higher concentration of americium inside sandy soil particles than in the water present in the soil pores; an even higher ratio was measured in loam soils. Americium is produced mostly artificially in small quantities, for research purposes. A tonne of spent nuclear fuel contains about 100 grams of various americium isotopes, mostly 241Am and 243Am. Their prolonged radioactivity is undesirable for the disposal, and therefore americium, together with other long-lived actinides, must be neutralized. The associated procedure may involve several steps, where americium is first separated and then converted by neutron bombardment in special reactors to short-lived nuclides. This procedure is well known as nuclear transmutation, but it is still being developed for americium. The transuranic elements from americium to fermium occurred naturally in the natural nuclear fission reactor at Oklo, but no longer do so. Americium is also one of the elements that have been detected in Przybylski's Star. Synthesis and extraction Isotope nucleosynthesis Americium has been produced in small quantities in nuclear reactors for decades, and kilograms of its 241Am and 243Am isotopes have been accumulated by now. Nevertheless, since it was first offered for sale in 1962, its price, about of 241Am, remains almost unchanged owing to the very complex separation procedure. The heavier isotope 243Am is produced in much smaller amounts; it is thus more difficult to separate, resulting in a higher cost of the order . Americium is not synthesized directly from uranium – the most common reactor material – but from the plutonium isotope 239Pu. The latter needs to be produced first, according to the following nuclear process: ^{238}_{92}U ->[\ce{(n,\gamma)}] ^{239}_{92}U ->[\beta^-][23.5 \ \ce{min}] ^{239}_{93}Np ->[\beta^-][2.3565 \ \ce{d}] ^{239}_{94}Pu The capture of two neutrons by 239Pu (a so-called (n,γ) reaction), followed by a β-decay, results in 241Am: ^{239}_{94}Pu ->[\ce{2(n,\gamma)}] ^{241}_{94}Pu ->[\beta^-][14.35 \ \ce{yr}] ^{241}_{95}Am The plutonium present in spent nuclear fuel contains about 12% of 241Pu. Because it beta-decays to 241Am, 241Pu can be extracted and may be used to generate further 241Am. However, this process is rather slow: half of the original amount of 241Pu decays to 241Am after about 15 years, and the 241Am amount reaches a maximum after 70 years. The obtained 241Am can be used for generating heavier americium isotopes by further neutron capture inside a nuclear reactor. In a light water reactor (LWR), 79% of 241Am converts to 242Am and 10% to its nuclear isomer 242mAm: Americium-242 has a half-life of only 16 hours, which makes its further conversion to 243Am extremely inefficient. The latter isotope is produced instead in a process where 239Pu captures four neutrons under high neutron flux: ^{239}_{94}Pu ->[\ce{4(n,\gamma)}] \ ^{243}_{94}Pu ->[\beta^-][4.956 \ \ce{h}] ^{243}_{95}Am Metal generation Most synthesis routines yield a mixture of different actinide isotopes in oxide forms, from which isotopes of americium can be separated. In a typical procedure, the spent reactor fuel (e.g. MOX fuel) is dissolved in nitric acid, and the bulk of uranium and plutonium is removed using a PUREX-type extraction (Plutonium–URanium EXtraction) with tributyl phosphate in a hydrocarbon. The lanthanides and remaining actinides are then separated from the aqueous residue (raffinate) by a diamide-based extraction, to give, after stripping, a mixture of trivalent actinides and lanthanides. Americium compounds are then selectively extracted using multi-step chromatographic and centrifugation techniques with an appropriate reagent. A large amount of work has been done on the solvent extraction of americium. For example, a 2003 EU-funded project codenamed "EUROPART" studied triazines and other compounds as potential extraction agents. A bis-triazinyl bipyridine complex was proposed in 2009 as such a reagent is highly selective to americium (and curium). Separation of americium from the highly similar curium can be achieved by treating a slurry of their hydroxides in aqueous sodium bicarbonate with ozone, at elevated temperatures. Both Am and Cm are mostly present in solutions in the +3 valence state; whereas curium remains unchanged, americium oxidizes to soluble Am(IV) complexes which can be washed away. Metallic americium is obtained by reduction from its compounds. Americium(III) fluoride was first used for this purpose. The reaction was conducted using elemental barium as reducing agent in a water- and oxygen-free environment inside an apparatus made of tantalum and tungsten. An alternative is the reduction of americium dioxide by metallic lanthanum or thorium: Physical properties In the periodic table, americium is located to the right of plutonium, to the left of curium, and below the lanthanide europium, with which it shares many physical and chemical properties. Americium is a highly radioactive element. When freshly prepared, it has a silvery-white metallic lustre, but then slowly tarnishes in air. With a density of 12 g/cm3, americium is less dense than both curium (13.52 g/cm3) and plutonium (19.8 g/cm3); but has a higher density than europium (5.264 g/cm3)—mostly because of its higher atomic mass. Americium is relatively soft and easily deformable and has a significantly lower bulk modulus than the actinides before it: Th, Pa, U, Np and Pu. Its melting point of 1173 °C is significantly higher than that of plutonium (639 °C) and europium (826 °C), but lower than for curium (1340 °C). At ambient conditions, americium is present in its most stable α form which has a hexagonal crystal symmetry, and a space group P63/mmc with cell parameters a = 346.8 pm and c = 1124 pm, and four atoms per unit cell. The crystal consists of a double-hexagonal close packing with the layer sequence ABAC and so is isotypic with α-lanthanum and several actinides such as α-curium. The crystal structure of americium changes with pressure and temperature. When compressed at room temperature to 5 GPa, α-Am transforms to the β modification, which has a face-centered cubic (fcc) symmetry, space group Fmm and lattice constant a = 489 pm. This fcc structure is equivalent to the closest packing with the sequence ABC. Upon further compression to 23 GPa, americium transforms to an orthorhombic γ-Am structure similar to that of α-uranium. There are no further transitions observed up to 52 GPa, except for an appearance of a monoclinic phase at pressures between 10 and 15 GPa. There is no consistency on the status of this phase in the literature, which also sometimes lists the α, β and γ phases as I, II and III. The β-γ transition is accompanied by a 6% decrease in the crystal volume; although theory also predicts a significant volume change for the α-β transition, it is not observed experimentally. The pressure of the α-β transition decreases with increasing temperature, and when α-americium is heated at ambient pressure, at 770 °C it changes into an fcc phase which is different from β-Am, and at 1075 °C it converts to a body-centered cubic structure. The pressure-temperature phase diagram of americium is thus rather similar to those of lanthanum, praseodymium and neodymium. As with many other actinides, self-damage of the crystal structure due to alpha-particle irradiation is intrinsic to americium. It is especially noticeable at low temperatures, where the mobility of the produced structure defects is relatively low, by broadening of X-ray diffraction peaks. This effect makes somewhat uncertain the temperature of americium and some of its properties, such as electrical resistivity. So for americium-241, the resistivity at 4.2 K increases with time from about 2 µOhm·cm to 10 µOhm·cm after 40 hours, and saturates at about 16 µOhm·cm after 140 hours. This effect is less pronounced at room temperature, due to annihilation of radiation defects; also heating to room temperature the sample which was kept for hours at low temperatures restores its resistivity. In fresh samples, the resistivity gradually increases with temperature from about 2 µOhm·cm at liquid helium to 69 µOhm·cm at room temperature; this behavior is similar to that of neptunium, uranium, thorium and protactinium, but is different from plutonium and curium which show a rapid rise up to 60 K followed by saturation. The room temperature value for americium is lower than that of neptunium, plutonium and curium, but higher than for uranium, thorium and protactinium. Americium is paramagnetic in a wide temperature range, from that of liquid helium, to room temperature and above. This behavior is markedly different from that of its neighbor curium which exhibits antiferromagnetic transition at 52 K. The thermal expansion coefficient of americium is slightly anisotropic and amounts to along the shorter a axis and for the longer c hexagonal axis. The enthalpy of dissolution of americium metal in hydrochloric acid at standard conditions is , from which the standard enthalpy change of formation (ΔfH°) of aqueous Am3+ ion is . The standard potential Am3+/Am0 is . Chemical properties Americium metal readily reacts with oxygen and dissolves in aqueous acids. The most stable oxidation state for americium is +3,. The chemistry of americium(III) has many similarities to the chemistry of lanthanide(III) compounds. For example, trivalent americium forms insoluble fluoride, oxalate, iodate, hydroxide, phosphate and other salts. Compounds of americium in oxidation states 2, 4, 5, 6 and 7 have also been studied. This is the widest range that has been observed with actinide elements. The color of americium compounds in aqueous solution is as follows: Am3+ (yellow-reddish), Am4+ (yellow-reddish), ; (yellow), (brown) and (dark green). The absorption spectra have sharp peaks, due to f-f transitions' in the visible and near-infrared regions. Typically, Am(III) has absorption maxima at ca. 504 and 811 nm, Am(V) at ca. 514 and 715 nm, and Am(VI) at ca. 666 and 992 nm. Americium compounds with oxidation state +4 and higher are strong oxidizing agents, comparable in strength to the permanganate ion () in acidic solutions. Whereas the Am4+ ions are unstable in solutions and readily convert to Am3+, compounds such as americium dioxide (AmO2) and americium(IV) fluoride (AmF4) are stable in the solid state. The pentavalent oxidation state of americium was first observed in 1951. In acidic aqueous solution the ion is unstable with respect to disproportionation. The reaction is typical. The chemistry of Am(V) and Am(VI) is comparable to the chemistry of uranium in those oxidation states. In particular, compounds like and are comparable to uranates and the ion is comparable to the uranyl ion, . Such compounds can be prepared by oxidation of Am(III) in dilute nitric acid with ammonium persulfate. Other oxidising agents that have been used include silver(I) oxide, ozone and sodium persulfate. Chemical compounds Oxygen compounds Three americium oxides are known, with the oxidation states +2 (AmO), +3 (Am2O3) and +4 (AmO2). Americium(II) oxide was prepared in minute amounts and has not been characterized in detail. Americium(III) oxide is a red-brown solid with a melting point of 2205 °C. Americium(IV) oxide is the main form of solid americium which is used in nearly all its applications. As most other actinide dioxides, it is a black solid with a cubic (fluorite) crystal structure. The oxalate of americium(III), vacuum dried at room temperature, has the chemical formula Am2(C2O4)3·7H2O. Upon heating in vacuum, it loses water at 240 °C and starts decomposing into AmO2 at 300 °C, the decomposition completes at about 470 °C. The initial oxalate dissolves in nitric acid with the maximum solubility of 0.25 g/L. Halides Halides of americium are known for the oxidation states +2, +3 and +4, where the +3 is most stable, especially in solutions. Reduction of Am(III) compounds with sodium amalgam yields Am(II) salts – the black halides AmCl2, AmBr2 and AmI2. They are very sensitive to oxygen and oxidize in water, releasing hydrogen and converting back to the Am(III) state. Specific lattice constants are: Orthorhombic AmCl2: a = , b = and c = Tetragonal AmBr2: a = and c = . They can also be prepared by reacting metallic americium with an appropriate mercury halide HgX2, where X = Cl, Br or I: {Am} + \underset{mercury\ halide}{HgX2} ->[{} \atop 400 - 500 ^\circ \ce C] {AmX2} + {Hg} Americium(III) fluoride (AmF3) is poorly soluble and precipitates upon reaction of Am3+ and fluoride ions in weak acidic solutions: Am^3+ + 3F^- -> AmF3(v) The tetravalent americium(IV) fluoride (AmF4) is obtained by reacting solid americium(III) fluoride with molecular fluorine: 2AmF3 + F2 -> 2AmF4 Another known form of solid tetravalent americium fluoride is KAmF5. Tetravalent americium has also been observed in the aqueous phase. For this purpose, black Am(OH)4 was dissolved in 15-M NH4F with the americium concentration of 0.01 M. The resulting reddish solution had a characteristic optical absorption spectrum which is similar to that of AmF4 but differed from other oxidation states of americium. Heating the Am(IV) solution to 90 °C did not result in its disproportionation or reduction, however a slow reduction was observed to Am(III) and assigned to self-irradiation of americium by alpha particles. Most americium(III) halides form hexagonal crystals with slight variation of the color and exact structure between the halogens. So, chloride (AmCl3) is reddish and has a structure isotypic to uranium(III) chloride (space group P63/m) and the melting point of 715 °C. The fluoride is isotypic to LaF3 (space group P63/mmc) and the iodide to BiI3 (space group R). The bromide is an exception with the orthorhombic PuBr3-type structure and space group Cmcm. Crystals of americium hexahydrate (AmCl3·6H2O) can be prepared by dissolving americium dioxide in hydrochloric acid and evaporating the liquid. Those crystals are hygroscopic and have yellow-reddish color and a monoclinic crystal structure. Oxyhalides of americium in the form AmVIO2X2, AmVO2X, AmIVOX2 and AmIIIOX can be obtained by reacting the corresponding americium halide with oxygen or Sb2O3, and AmOCl can also be produced by vapor phase hydrolysis: AmCl3 + H2O -> AmOCl + 2HCl Chalcogenides and pnictides The known chalcogenides of americium include the sulfide AmS2, selenides AmSe2 and Am3Se4, and tellurides Am2Te3 and AmTe2. The pnictides of americium (243Am) of the AmX type are known for the elements phosphorus, arsenic, antimony and bismuth. They crystallize in the rock-salt lattice. Silicides and borides Americium monosilicide (AmSi) and "disilicide" (nominally AmSix with: 1.87 < x < 2.0) were obtained by reduction of americium(III) fluoride with elementary silicon in vacuum at 1050 °C (AmSi) and 1150−1200 °C (AmSix). AmSi is a black solid isomorphic with LaSi, it has an orthorhombic crystal symmetry. AmSix has a bright silvery lustre and a tetragonal crystal lattice (space group I41/amd), it is isomorphic with PuSi2 and ThSi2. Borides of americium include AmB4 and AmB6. The tetraboride can be obtained by heating an oxide or halide of americium with magnesium diboride in vacuum or inert atmosphere. Organoamericium compounds Analogous to uranocene, americium forms the organometallic compound amerocene with two cyclooctatetraene ligands, with the chemical formula (η8-C8H8)2Am. A cyclopentadienyl complex is also known that is likely to be stoichiometrically AmCp3. Formation of the complexes of the type Am(n-C3H7-BTP)3, where BTP stands for 2,6-di(1,2,4-triazin-3-yl)pyridine, in solutions containing n-C3H7-BTP and Am3+ ions has been confirmed by EXAFS. Some of these BTP-type complexes selectively interact with americium and therefore are useful in its selective separation from lanthanides and another actinides. Biological aspects Americium is an artificial element of recent origin, and thus does not have a biological requirement. It is harmful to life. It has been proposed to use bacteria for removal of americium and other heavy metals from rivers and streams. Thus, Enterobacteriaceae of the genus Citrobacter precipitate americium ions from aqueous solutions, binding them into a metal-phosphate complex at their cell walls. Several studies have been reported on the biosorption and bioaccumulation of americium by bacteria and fungi. Fission The isotope 242mAm (half-life 141 years) has the largest cross sections for absorption of thermal neutrons (5,700 barns), that results in a small critical mass for a sustained nuclear chain reaction. The critical mass for a bare 242mAm sphere is about 9–14 kg (the uncertainty results from insufficient knowledge of its material properties). It can be lowered to 3–5 kg with a metal reflector and should become even smaller with a water reflector. Such small critical mass is favorable for portable nuclear weapons, but those based on 242mAm are not known yet, probably because of its scarcity and high price. The critical masses of two other readily available isotopes, 241Am and 243Am, are relatively high – 57.6 to 75.6 kg for 241Am and 209 kg for 243Am. Scarcity and high price yet hinder application of americium as a nuclear fuel in nuclear reactors. There are proposals of very compact 10-kW high-flux reactors using as little as 20 grams of 242mAm. Such low-power reactors would be relatively safe to use as neutron sources for radiation therapy in hospitals. Isotopes About 19 isotopes and 8 nuclear isomers are known for americium. There are two long-lived alpha-emitters; 243Am has a half-life of 7,370 years and is the most stable isotope, and 241Am has a half-life of 432.2 years. The most stable nuclear isomer is 242m1Am; it has a long half-life of 141 years. The half-lives of other isotopes and isomers range from 0.64 microseconds for 245m1Am to 50.8 hours for 240Am. As with most other actinides, the isotopes of americium with odd number of neutrons have relatively high rate of nuclear fission and low critical mass. Americium-241 decays to 237Np emitting alpha particles of 5 different energies, mostly at 5.486 MeV (85.2%) and 5.443 MeV (12.8%). Because many of the resulting states are metastable, they also emit gamma rays with the discrete energies between 26.3 and 158.5 keV. Americium-242 is a short-lived isotope with a half-life of 16.02 h. It mostly (82.7%) converts by β-decay to 242Cm, but also by electron capture to 242Pu (17.3%). Both 242Cm and 242Pu transform via nearly the same decay chain through 238Pu down to 234U. Nearly all (99.541%) of 242m1Am decays by internal conversion to 242Am and the remaining 0.459% by α-decay to 238Np. The latter subsequently decays to 238Pu and then to 234U. Americium-243 transforms by α-emission into 239Np, which converts by β-decay to 239Pu, and the 239Pu changes into 235U by emitting an α-particle. Applications Ionization-type smoke detector Americium is used in the most common type of household smoke detector, which uses 241Am in the form of americium dioxide as its source of ionizing radiation. This isotope is preferred over 226Ra because it emits 5 times more alpha particles and relatively little harmful gamma radiation. The amount of americium in a typical new smoke detector is 1 microcurie (37 kBq) or 0.29 microgram. This amount declines slowly as the americium decays into neptunium-237, a different transuranic element with a much longer half-life (about 2.14 million years). With its half-life of 432.2 years, the americium in a smoke detector includes about 3% neptunium after 19 years, and about 5% after 32 years. The radiation passes through an ionization chamber, an air-filled space between two electrodes, and permits a small, constant current between the electrodes. Any smoke that enters the chamber absorbs the alpha particles, which reduces the ionization and affects this current, triggering the alarm. Compared to the alternative optical smoke detector, the ionization smoke detector is cheaper and can detect particles which are too small to produce significant light scattering; however, it is more prone to false alarms. Radionuclide As 241Am has a roughly similar half-life to 238Pu (432.2 years vs. 87 years), it has been proposed as an active element of radioisotope thermoelectric generators, for example in spacecraft. Although americium produces less heat and electricity – the power yield is 114.7 mW/g for 241Am and 6.31 mW/g for 243Am (cf. 390 mW/g for 238Pu) – and its radiation poses more threat to humans owing to neutron emission, the European Space Agency is considering using americium for its space probes. Another proposed space-related application of americium is a fuel for space ships with nuclear propulsion. It relies on the very high rate of nuclear fission of 242mAm, which can be maintained even in a micrometer-thick foil. Small thickness avoids the problem of self-absorption of emitted radiation. This problem is pertinent to uranium or plutonium rods, in which only surface layers provide alpha-particles. The fission products of 242mAm can either directly propel the spaceship or they can heat a thrusting gas. They can also transfer their energy to a fluid and generate electricity through a magnetohydrodynamic generator. One more proposal which utilizes the high nuclear fission rate of 242mAm is a nuclear battery. Its design relies not on the energy of the emitted by americium alpha particles, but on their charge, that is the americium acts as the self-sustaining "cathode". A single 3.2 kg 242mAm charge of such battery could provide about 140 kW of power over a period of 80 days. Even with all the potential benefits, the current applications of 242mAm are as yet hindered by the scarcity and high price of this particular nuclear isomer. In 2019, researchers at the UK National Nuclear Laboratory and the University of Leicester demonstrated the use of heat generated by americium to illuminate a small light bulb. This technology could lead to systems to power missions with durations up to 400 years into interstellar space, where solar panels do not function. Neutron source The oxide of 241Am pressed with beryllium is an efficient neutron source. Here americium acts as the alpha source, and beryllium produces neutrons owing to its large cross-section for the (α,n) nuclear reaction: ^{241}_{95}Am -> ^{237}_{93}Np + ^{4}_{2}He + \gamma ^{9}_{4}Be + ^{4}_{2}He -> ^{12}_{6}C + ^{1}_{0}n + \gamma The most widespread use of 241AmBe neutron sources is a neutron probe – a device used to measure the quantity of water present in soil, as well as moisture/density for quality control in highway construction. 241Am neutron sources are also used in well logging applications, as well as in neutron radiography, tomography and other radiochemical investigations. Production of other elements Americium is a starting material for the production of other transuranic elements and transactinides – for example, 82.7% of 242Am decays to 242Cm and 17.3% to 242Pu. In the nuclear reactor, 242Am is also up-converted by neutron capture to 243Am and 244Am, which transforms by β-decay to 244Cm: ^{243}_{95}Am ->[\ce{(n,\gamma)}] ^{244}_{95}Am ->[\beta^-][10.1 \ \ce{h}] ^{244}_{96}Cm Irradiation of 241Am by 12C or 22Ne ions yields the isotopes 247Es (einsteinium) or 260Db (dubnium), respectively. Furthermore, the element berkelium (243Bk isotope) had been first intentionally produced and identified by bombarding 241Am with alpha particles, in 1949, by the same Berkeley group, using the same 60-inch cyclotron. Similarly, nobelium was produced at the Joint Institute for Nuclear Research, Dubna, Russia, in 1965 in several reactions, one of which included irradiation of 243Am with 15N ions. Besides, one of the synthesis reactions for lawrencium, discovered by scientists at Berkeley and Dubna, included bombardment of 243Am with 18O. Spectrometer Americium-241 has been used as a portable source of both gamma rays and alpha particles for a number of medical and industrial uses. The 59.5409 keV gamma ray emissions from 241Am in such sources can be used for indirect analysis of materials in radiography and X-ray fluorescence spectroscopy, as well as for quality control in fixed nuclear density gauges and nuclear densometers. For example, the element has been employed to gauge glass thickness to help create flat glass. Americium-241 is also suitable for calibration of gamma-ray spectrometers in the low-energy range, since its spectrum consists of nearly a single peak and negligible Compton continuum (at least three orders of magnitude lower intensity). Americium-241 gamma rays were also used to provide passive diagnosis of thyroid function. This medical application is however obsolete. Health concerns As a highly radioactive element, americium and its compounds must be handled only in an appropriate laboratory under special arrangements. Although most americium isotopes predominantly emit alpha particles which can be blocked by thin layers of common materials, many of the daughter products emit gamma-rays and neutrons which have a long penetration depth. If consumed, most of the americium is excreted within a few days, with only 0.05% absorbed in the blood, of which roughly 45% goes to the liver and 45% to the bones, and the remaining 10% is excreted. The uptake to the liver depends on the individual and increases with age. In the bones, americium is first deposited over cortical and trabecular surfaces and slowly redistributes over the bone with time. The biological half-life of 241Am is 50 years in the bones and 20 years in the liver, whereas in the gonads (testicles and ovaries) it remains permanently; in all these organs, americium promotes formation of cancer cells as a result of its radioactivity. Americium often enters landfills from discarded smoke detectors. The rules associated with the disposal of smoke detectors are relaxed in most jurisdictions. In 1994, 17-year-old David Hahn extracted the americium from about 100 smoke detectors in an attempt to build a breeder nuclear reactor. There have been a few cases of exposure to americium, the worst case being that of chemical operations technician Harold McCluskey, who at the age of 64 was exposed to 500 times the occupational standard for americium-241 as a result of an explosion in his lab. McCluskey died at the age of 75 of unrelated pre-existing disease. See also Actinides in the environment :Category:Americium compounds Notes References Bibliography Penneman, R. A. and Keenan T. K. The radiochemistry of americium and curium, University of California, Los Alamos, California, 1960 Further reading Nuclides and Isotopes – 14th Edition, GE Nuclear Energy, 1989. External links Americium at The Periodic Table of Videos (University of Nottingham) ATSDR – Public Health Statement: Americium World Nuclear Association – Smoke Detectors and Americium Chemical elements Chemical elements with double hexagonal close-packed structure Actinides Carcinogens Synthetic elements
1087
https://en.wikipedia.org/wiki/Foreign%20relations%20of%20Azerbaijan
Foreign relations of Azerbaijan
The Republic of Azerbaijan is a member of the United Nations, the Non-Aligned Movement, the Organization for Security and Cooperation in Europe, NATO's Partnership for Peace, the Euro-Atlantic Partnership Council, the World Health Organization, the European Bank for Reconstruction and Development; the Council of Europe, CFE Treaty, the Community of Democracies; the International Monetary Fund; and the World Bank. The major trends in the foreign relations of the Republic of Azerbaijan toward both global and regional powers active in Caucasus area. External variables are categorized depending on their original nature into two groups: global and regional. The former category includes global players such as Moscow and Washington, while the latter category rival regional players, namely Ankara and Tehran. Azerbaijan has formal involvement with senior ex-U.S. government officials including James Baker and Henry Kissinger, as they serve on the Honorary Council of Advisors of the U.S.-Azerbaijan Chamber of Commerce (USACC). USACC is co-chaired by Tim Cejka, President of ExxonMobil and Reza Vaziri, President of R.V. Investment Group and Chairman of the Anglo Asian Mining Plc (LSE Ticker: AAZ). Diplomatic relations As of 2023, Azerbaijan maintains diplomatic relations with 184 United Nations member states, the State of Palestine and the Holy See. Azerbaijan does not have diplomatic relations with the following countries: Armenia Botswana Central African Republic Cyprus Republic of the Congo Kiribati Federated States of Micronesia Tonga Azerbaijan also maintains good relations with the European Union, in the framework of its Eastern European Neighbourhood Policy (See Azerbaijan and the European Union). List List of countries with which Azerbaijan maintains diplomatic relations with: Information on some of the countries with which Azerbaijan maintains formal relations Multilateral Africa Americas Asia Europe Oceania International organizations AsDB BSEC CE CIS EAPC EBRD ECE ECO ESCAP FAO GUAM IAEA IBRD ICAO ICRM IDA IDB IFAD IFC IFRCS ILO IMF IMO Interpol IOC, IOM ISO (correspondent) ITU ITUC OAS (observer) OIC OPCW OSCE PFP United Nations UNCTAD UNESCO UNIDO UPU WCO WFTU WHO WIPO WMO WToO WTrO(observer) Other entities Sovereign Military Order of Malta – there are no relations States with limited recognition Disputes Nagorno-Karabakh/Azerbaijan The frozen conflict over currently largely Armenian-populated region of Nagorno-Karabakh within the Republic of Azerbaijan began when in 1988 the Armenian majority of Nagorno-Karabakh demanded autonomy with demonstrations and persecutions against ethnic Azeris following in Armenia. This led to anti-Armenian rioting in Azerbaijan, with Azerbaijani militias beginning their effort to expel Armenians from the enclave. In 1992, a war broke out and pogroms of Armenians and Azeris forced both groups to flee their homes. In 1994, a Russian-brokered ceasefire ended the war but more than 1 million ethnic Armenians and Azeris are still not able to return home. The conflict over Nagorno-Karabakh remains unresolved despite negotiations, that are ongoing since 1992 under the aegis of the Minsk Group of the OSCE, to resolve the conflict peacefully. Caviar diplomacy The European Stability Initiative (ESI) has revealed in a report from 2012 with the title "Caviar diplomacy: How Azerbaijan silenced the Council of Europe", that since Azerbaijan's entry into the Council of Europe, each year 30 to 40 deputies are invited to Azerbaijan and generously paid with expensive gifts, including caviar (worth up to 1,400 euro), silk carpets, gold, silver and large amounts of money. In return they become lobbyists for Azerbaijan. This practice has been widely referred to as "Caviar diplomacy". ESI also published a report on 2013 Presidential elections in Azerbaijan titled "Disgraced: Azerbaijan and the end of election monitoring as we know it". The report revealed the ties between Azerbaijani government and the members of certain observation missions who praised the elections. Azerbaijan's "Caviar diplomacy" at 2013 presidential elections sparked a major international scandal, as the reports of two authoritative organizations Parliamentary Assembly of the Council of Europe/European Parliament and OSCE/ODIHR completely contradicted one another in their assessments of elections. Non-governmental anti-corruption organization Transparency International has regularly judged Azerbaijan to be one of the most corrupt countries in the world and has also criticized Azerbaijan for the "Caviar diplomacy". At June 2016 the public prosecutor of Milan has accused the former leader of the (Christian) Union of the center and of the European People's Party of the Parliamentary Assembly of the Council of Europe Luca Volonte of accepting large bribes from representatives of the Azerbaijani government. Two people with high-level experience of the Council of Europe's parliamentary assembly (Pace) have told the Guardian they believe its members have been offered bribes for votes by Azerbaijan. Former Azerbaijani diplomat, Arif Mammadov, alleged that a member of Azerbaijan's delegation at the Council of Europe had €30m (£25m) to spend on lobbying its institutions, including the Council of Europe assembly. PACE ratified the terms of reference of an independent external investigation body to carry out a detailed independent inquiry into the allegations of corruption at the council involving Azerbaijan. ESISC report On 6 March 2017, ESISC (European Strategic Intelligence and Security Center) published a scandalous report called "The Armenian Connection" where it veraciously attacked human rights NGOs and research organisations criticising human rights violations and corruption in Azerbaijan, Turkey, and Russia. ESISC in that report asserted that "Caviar diplomacy" report elaborated by ESI aimed to create climate of suspicion based on slander to form a network of MPs that would engage in a political war against Azerbaijan. In the Second Chapter of the report called "The Armenian Connection: «Mr X», Nils Muižnieks, Council of Europe Commissioner for Human Rights" that was published on 18 April 2017 ESISC asserted that the network composed of European PMs, Armenian officials and some NGOs: Human Rights Watch, Amnesty International, "Human Rights House Foundation", "Open Dialog", European Stability Initiative, and Helsinki Committee for Human Rights, was financed by the Soros Foundation. According to ESISC the key figure of the network since 2012 has been Nils Muižnieks, Commissioner for Human Rights of the Council of Europe and the network has served to the interests of George Soros and the Republic of Armenia. "The report is written in the worst traditions of authoritarian propaganda, makes absurd claims, and is clearly aimed at deflecting the wave of criticism against cover-up of unethical lobbying and corruption in PACE and demands for change in the Assembly", said Freedom Files Analytical Centre. According Robert Coalson (Radio Free Europe), ESISC is a part of Baku's lobbying efforts to extend to the use of front think tanks to shift public opinion. European Stability Initiative said that "ESISC report is full of lies (such as claiming that German PACE member Strasser holds pro-Armenian views and citing as evidence that he went to Yerevan in 2015 to commemorate the Armenian genocide, when Strasser has never in his life been to independent Armenia)". See also Azerbaijan and the European Union Azerbaijan and the International Monetary Fund List of diplomatic missions in Azerbaijan List of diplomatic missions of Azerbaijan Visa requirements for Azerbaijani citizens Further reading Valiyev, Anar: "Azerbaijan and the North Caucasus: A Pragmatic Relationship" in the Caucasus Analytical Digest No. 27 Hübner, Gerald: "Foreign Direct Investment in Azerbaijan—the Quality of Quantity" in the Caucasus Analytical Digest No. 28 Abbasov, Shahin: "Azerbaijan's Eurovision Story: Great Chances to Improve, But No Political Will" in the Caucasus Analytical Digest No. 32 Mazziotti, Marius; Sauerborn, Djan; Scianna, Bastian Matteo: "Multipolarity is key: Assessing Azerbaijan's foreign policy" References CIA World Factbook 2000 and the 2003 U.S. Department of State website External links U.S. Embassy in Azerbaijan in Baku Embassy of the Republic of Azerbaijan in Washington Tajikistan & Kyrgyzstan relations
1270
https://en.wikipedia.org/wiki/Extreme%20poverty
Extreme poverty
Extreme poverty is the most severe type of poverty, defined by the United Nations (UN) as "a condition characterized by severe deprivation of basic human needs, including food, safe drinking water, sanitation facilities, health, shelter, education and information. It depends not only on income but also on access to services". Historically, other definitions have been proposed within the United Nations. In 2018, extreme poverty mainly refers to an income below the international poverty line of $1.90 per day (in 2011 prices, $ in dollars), set by the World Bank. In October 2017, the World Bank updated the international poverty line, a global absolute minimum, to $1.90 a day. This is the equivalent of $1.00 a day in 1996 US prices, hence the widely used expression "living on less than a dollar a day". The vast majority of those in extreme poverty reside in South Asia and Sub-Saharan Africa. As of 2018, it is estimated that the country with the most people living in extreme poverty is Nigeria, at 86 million. In the past, the vast majority of the world population lived in conditions of extreme poverty. The percentage of the global population living in absolute poverty fell from over 80% in 1800 to under 20% by 2015. According to UN estimates, roughly 734 million people or 10% remained under those conditions. The number had previously been measured as 1.9 billion in 1990, and 1.2 billion in 2008. Despite the significant number of individuals still below the international poverty line, these figures represent significant progress for the international community, as they reflect a decrease of more than one billion people over 15 years. In public opinion surveys around the globe, people surveyed tend to think that extreme poverty has not decreased. The reduction of extreme poverty and hunger was the first Millennium Development Goal (MDG1), as set by the United Nations in 2000. Specifically, the target was to reduce the extreme poverty rate by half by 2015, a goal that was met five years ahead of schedule. In the Sustainable Development Goals, which succeeded the MDGs, the goal is to end extreme poverty in all its forms everywhere. With this declaration the international community, including the UN and the World Bank have adopted the target of ending extreme poverty by 2030. Definition Previous definitions In July 1993, Leandro Despouy, the then UN Special Rapporteur on extreme poverty and human rights made use of a definition he adapted from a 1987 report to the French Economic and Social Council by Fr. Joseph Wresinski, founder of the International Movement ATD Fourth World, distinguishing "lack of basic security" (poverty) and "chronic poverty" (extreme poverty), linking the eradication of extreme poverty by allowing people currently experiencing it a real opportunity to exercise all their human rights: "The lack of basic security connotes the absence of one or more factors enabling individuals and families to assume basic responsibilities and to enjoy fundamental rights. The situation may become widespread and result in more serious and permanent consequences. The lack of basic security leads to chronic poverty when it simultaneously affects several aspects of people's lives, when it is prolonged and when it severely compromises people's chances of regaining their rights and of reassuming their responsibilities in the foreseeable future." This definition was mentioned previously, in June 1989, in the preliminary report on the realization of economic, social and cultural rights by the UN Special Rapporteur Danilo Türk. It is still in use today, among others, in the current UN Guiding Principles on Extreme Poverty and Human Rights adopted by the UN Human Rights Council in September 2012 Consumption-based definition Extreme poverty is defined by the international community as living below $1.90 a day, as measured in 2011 international prices (equivalent to $2.12 in 2018). This number, also known as the international poverty line, is periodically updated to account for inflation and differences in the cost of living; it was originally defined at $1.00 a day in 1996. The updates are made according to new price data to portray the costs of basic food, health services, clothing, and shelter around the world as accurately as possible. The latest revision was made in 2015 when the World Bank increased the line to international- $1.90. Because many of the world's poorest people do not have a monetary income, the poverty measurement is based on the monetary value of a person's consumption. Otherwise the poverty measurement would be missing the home production of subsistence farmers that consume largely their own production. Alternative definitions The $1.90/day extreme poverty line remains the most widely used metric as it highlights the reality of those in the most severe conditions. Although widely used by most international organizations, it has come under scrutiny due to a variety of factors. For example, it does not account for how far below the line people are, referred to as the depth of poverty. For this purpose, the same institutions publish data on the poverty gap. The international poverty line is designed to stay constant over time, to allow comparisons between different years. It is therefore a measure of absolute poverty and is not measuring relative poverty. It is also not designed to capture how people view their own financial situation (known as the socially subjective poverty line). Moreover, the calculation of the poverty line relies on information about consumer prices to calculate purchasing power parity, which are very hard to measure and are necessarily debatable. As with all other metrics, there may also be missing data from the poorest and most fragile countries. Several alternative instruments for measuring extreme poverty have been suggested which incorporate other factors such as malnutrition and lack of access to a basic education. The Multidimensional Poverty Index (MPI), based on the Alkire-Foster Method, is published by the Oxford Poverty & Human Development Initiative (OPHI): it measures deprivation in basic needs and can be broken down to reflect both the incidence and the intensity of poverty. For example, under conventional measures, in both Ethiopia and Uzbekistan about 40% of the population is considered extremely poor, but based on the MPI, 90% of Ethiopians but only 2% of Uzbeks are in multidimensional poverty. The MPI is useful for development officials to determine the most likely causes of poverty within a region, using the M0 measure of the method (which is calculated by multiplying the fraction of people in poverty by the fraction of dimensions they are deprived in). For example, in the Gaza Strip of Palestine, using the M0 measure of the Alkire-Foster method reveals that poverty in the region is primarily caused by a lack of access to electricity, lack of access to drinking water, and widespread overcrowding. In contrast, data from the Chhukha District of Bhutan reveals that income is a much larger contributor to poverty as opposed to other dimensions within the region. However, the MPI only presents data from 105 countries, so it cannot be used for global measurements. Share of the population living in extreme poverty Current trends Getting to zero Using the World Bank definition of $1.90/day, , roughly 710 million people remained in extreme poverty (or roughly 1 in 10 people worldwide). Nearly half of them live in India and China, with more than 85% living in just 20 countries. Since the mid-1990s, there has been a steady decline in both the worldwide poverty rate and the total number of extreme poor. In 1990, the percentage of the global population living in extreme poverty was 43%, but in 2011, that percentage had dropped down to 21%. This halving of the extreme poverty rate falls in line with the first Millennium Development Goal (MDG1) proposed by former UN Secretary-General Kofi Annan, who called on the international community at the turn of the century to reduce the percentage of people in extreme poverty by half by 2015. This reduction in extreme poverty took place most notably in China, Indonesia, India, Pakistan and Vietnam. These five countries accounted for the alleviation of 715 million people out of extreme poverty between 1990 and 2010 – more than the global net total of roughly 700 million. This statistical oddity can be explained by the fact that the number of people living in extreme poverty in Sub-Saharan Africa rose from 290 million to 414 million over the same period. However, there have been many positive signs for extensive, global poverty reduction as well. Since 1999, the total number of extreme poor has declined by an average of 50 million per year. Moreover, in 2005, for the first time in recorded history, poverty rates began to fall in every region of the world, including Africa. As aforementioned, the number of people living in extreme poverty has reduced from 1.9 billion to 766 million over the span of the last decades. If we remain on our current trajectory, many economists predict we could reach global zero by 2030–2035, thus ending extreme poverty. Global zero entails a world in which fewer than 3% of the global population lives in extreme poverty (projected under most optimistic scenarios to be fewer than 200 million people). This zero figure is set at 3% in recognition of the fact that some amount of frictional (temporary) poverty will continue to exist, whether it is caused by political conflict or unexpected economic fluctuations, at least for the foreseeable future. However, the Brookings Institution notes that any projection about poverty more than a few years into the future runs the risk of being highly uncertain. This is because changes in consumption and distribution throughout the developing world over the next two decades could result in monumental shifts in global poverty, for better or worse. Others are more pessimistic about this possibility, predicting a range of 193 million to 660 million people still living in extreme poverty by 2035. Additionally, some believe the rate of poverty reduction will slow down in the developing world, especially in Africa, and as such it will take closer to five decades to reach global zero. Despite these reservations, several prominent international and national organizations, including the UN, the World Bank and the United States Federal Government (via USAID), have set a target of reaching global zero by the end of 2030. Exacerbating factors There are a variety of factors that may reinforce or instigate the existence of extreme poverty, such as weak institutions, cycles of violence and a low level of growth. Recent World Bank research shows that some countries can get caught in a "fragility trap", in which self-reinforcing factors prevent the poorest nations from emerging from low-level equilibrium in the long run. Moreover, most of the reduction in extreme poverty over the past twenty years has taken place in countries that have not experienced a civil conflict or have had governing institutions with a strong capacity to actually govern. Thus, to end extreme poverty, it is also important to focus on the interrelated problems of fragility and conflict. USAID defines fragility as a government's lack of both legitimacy (the perception the government is adequate at doing its job) and effectiveness (how good the government is at maintaining law and order, in an equitable manner). As fragile nations are unable to equitably and effectively perform the functions of a state, these countries are much more prone to violent unrest and mass inequality. Additionally, in countries with high levels of inequality (a common problem in countries with inadequate governing institutions), much higher growth rates are needed to reduce the rate of poverty when compared with other nations. Additionally, if China and India are removed from the equation, up to 70% of the world's poor live in fragile states by some definitions of fragility. Some analysts project that extreme poverty will be increasingly concentrated in fragile, low-income states like Haiti, Yemen and the Central African Republic. However, some academics, such as Andy Sumner, say that extreme poverty will be increasingly concentrated in middle-income countries, creating a paradox where the world's poor do not actually live in the poorest countries. To help low-income earners, fragile states make the transition towards peace and prosperity, the New Deal for Engagement in Fragile States, endorsed by roughly forty countries and multilateral institutions, was created in 2011. This represents an important step towards redressing the problem of fragility as it was originally articulated by self-identified fragile states who called on the international community to not only "do things differently", but to also "do different things". Civil conflict also remains a prime cause for the perpetuation of poverty throughout the developing world. Armed conflict can have severe effects on economic growth for many reasons such as the destruction of assets, destruction of livelihoods, creation of unwanted mass migration, and diversion of public resources towards war. Significantly, a country that experienced major violence during 1981–2005 had extreme poverty rates 21 percentage points higher than a country with no violence. On average, each civil conflict will cost a country roughly 30 years of GDP growth. Therefore, a renewed commitment from the international community to address the deteriorating situation in highly fragile states is necessary to both prevent the mass loss of life, but to also prevent the vicious cycle of extreme poverty. Population trends and dynamics (e.g. population growth) can also have a large impact on prospects for poverty reduction. According to the United Nations, "in addition to improving general health and well-being, analysis shows that meeting the reproductive health and contraceptive needs of all women in the developing world more than pays for itself"). In 2013, a prevalent finding in a report by the World Bank was that extreme poverty is most prevalent in low-income countries. In these countries, the World Bank found that progress in poverty reduction is the slowest, the poor live under the worst conditions, and the most affected persons are children age 12 and under. International initiatives Millennium Summit and Millennium Development Goals In September 2000, world leaders gathered at the Millennium Summit held in New York, launching the United Nations Millennium Project suggested by then UN Secretary-General Kofi Annan. Prior to the launch of the conference, the office of Secretary-General Annan released a report entitled "We The Peoples: The Role of the United Nations in the 21st Century". In this document, now widely known as the Millennium Report, Kofi Annan called on the international community to reduce the proportion of people in extreme poverty by half by 2015, a target that would affect over 1 billion people. Citing the close correlation between economic growth and the reduction of poverty in poor countries, Annan urged international leaders to indiscriminately target the problem of extreme poverty across every region. In charge of managing the project was Jeffrey Sachs, a noted development economist, who in 2005 released a plan for action called "Investing in Development: A Practical Plan to Achieve the Millennium Development Goals". Thomas Pogge criticized the 2000 Millennium Declaration for being less ambitious than a previous declaration from the World Food Summit due to using 1990 as the benchmark rather than 1996. Overall, there has been significant progress towards reducing extreme poverty, with the MDG1 target of reducing extreme poverty rates by half being met five years early, representing 700 million people being lifted out of extreme poverty from 1990 to 2010, with 1.2 billion people still remaining under those conditions. The notable exception to this trend was in Sub-Saharan Africa, the only region where the number of people living in extreme poverty rose from 290 million in 1990 to 414 million in 2010, comprising more than a third of those living in extreme poverty worldwide. 2005 World Summit The 2005 World Summit, held in September which was organized to measure international progress towards fulfilling the Millennium Development Goals (MDGs). Notably, the conference brought together more than 170 Heads of State. While world leaders at the summit were encouraged by the reduction of poverty in some nations, they were concerned by the uneven decline of poverty within and among different regions of the globe. However, at the end of the summit, the conference attendees reaffirmed the UN's commitment to achieve the MDGs by 2015 and urged all supranational, national and non-governmental organizations to follow suit. Sustainable Development Goals As the expiration of the Millennium Development Goals approached in 2015, the UN convened a panel to advise on a Post-2015 Development Agenda, which led to a new set of 17 goals for 2030 titled the Sustainable Development Goals (SDGs). The first goal (SDG 1) is to "End poverty in all its forms everywhere." The HLP report, entitled A New Global Partnership: Eradicate Poverty and Transform Economies Through Sustainable Development, was published in May 2013. In the report, the HLP wrote that: Ending extreme poverty is just the beginning, not the end. It is vital, but our vision must be broader: to start countries on the path of sustainable development – building on the foundations established by the 2012 UN Conference on Sustainable Development in Rio de Janeiro, and meeting a challenge that no country, developed or developing, has met so far. We recommend to the Secretary-General that deliberations on a new development agenda must be guided by the vision of eradicating extreme poverty once and for all, in the context of sustainable development. Therefore, the report determined that a central goal of the Post-Millennium Development agenda is to eradicate extreme poverty by 2030. However, the report also emphasized that the MDGs were not enough on their own, as they did not "focus on the devastating effects of conflict and violence on development ... the importance to development of good governance and institution ... nor the need for inclusive growth..." Consequently, there now exists synergy between the policy position papers put forward by the United States (through USAID), the World Bank and the UN itself in terms of viewing fragility and a lack of good governance as exacerbating extreme poverty. However, in a departure from the views of other organizations, the commission also proposed that the UN focus not only on extreme poverty (a line drawn at $1.25), but also on a higher target, such as $2. The report notes this change could be made to reflect the fact that escaping extreme poverty is only a first step. In addition to the UN, a host of other supranational and national actors such as the European Union and the African Union have published their own positions or recommendations on what should be incorporated in the Post-2015 agenda. The European Commission's communication, published in A decent Life for all: from vision to collective action, affirmed the UN's commitment to "eradicate extreme poverty in our lifetime and put the world on a sustainable path to ensure a decent life for all by 2030". A unique vision of the report was the commission's environmental focus (in addition to a plethora of other goals such as combating hunger and gender inequality). Specifically, the Commission argued, "long-term poverty reduction ... requires inclusive and sustainable growth. Growth should create decent jobs, take place with resource efficiency and within planetary boundaries, and should support efforts to mitigate climate change." The African Union's report, entitled Common African Position (CAP) on the Post-2015 Development Agenda, likewise encouraged the international community to focus on eradicating the twin problems of poverty and exclusion in our lifetime. Moreover, the CAP pledged that "no person – regardless of ethnicity, gender, geography, disability, race or other status – is denied universal human rights and basic economic opportunities". Least developed country conferences The UN least developed country (LDC) conferences were a series of summits organized by the UN to promote the substantial and even development of the world's least developed countries. The first UN LDC Conference was held between 1 and 14 September 1981, in Paris, the first UN LDC Conference was organized to finalize the UN's "Substantial New Programme of Action" for the 1980s in Least Developed Countries. This program, which was unanimously adopted by the conference attendees, argued for internal reforms in LDCs (meant to encourage economic growth) to be complemented by strong international measures. However, despite the major economic and policy reforms initiated many of these LDCs, in addition to strong international aid, the economic situation of these countries worsened as a whole in the 1980s. This prompted the organization of a 2nd UN LDC conference almost a decade later. The second UN LDC Conference was held between 3 and 14 September 1990, once again in Paris, the second UN LDC Conference was convened to measure the progress made by the LDCs towards fulfilling their development goals during the 1980s. Recognizing the problems that plagued the LDCs over the past decade, the conference formulated a new set of national and international policies to accelerate the growth rates of the poorest nations. These new principles were embodied in the "Paris Declaration and Programme of Action for the Least Developed Countries for the 1990s". The fourth UN LDC Conference was the most recent conference. It was held in May 2011 in Istanbul, recognized that the nature of development had fundamentally changed since the 1st conference held almost 30 years earlier. In the 21st century, the capital flow into emerging economies has increasingly become dominated by foreign direct investment and remittances, as opposed to bilateral and multilateral assistance. Moreover, since the 1980s, significant structural changes have taken place on the international stage. With the creation of the G-20 conference of the largest economic powers, including many nations in the Global South, formerly undeveloped nations are now able to have a much larger say in international relations. Furthermore, the conference recognized that in the midst of a deep global recession, coupled with multiple crises (energy, climate, food, etc.), the international community would have fewer resources to aid the LDCs. Thus, the UN considered the participation of a wide range of stakeholders (not least the LDCs themselves), crucial to the formulation of the conference. Organizations working to end extreme poverty International organizations World Bank In 2013, the Board of Governors of the World Bank Group (WBG) set two overriding goals for the WBG to commit itself to in the future. First, to end extreme poverty by 2030, an objective that echoes the sentiments of the UN and the Obama administration. Additionally, the WBG set an interim target of reducing extreme poverty to below 9% by 2020. Second, to focus on growth among the bottom 40% of people, as opposed to standard GDP growth. This commitment ensures that the growth of the developing world lifts people out of poverty, rather than exacerbating inequality. As the World Bank's primary focus is on delivering economic growth to enable equitable prosperity, its developments programs are primarily commercial-based in nature, as opposed to the UN. Since the World Bank recognizes better jobs will result in higher income, and thus less poverty, the WBG seeks to support employment training initiatives, small business development programs and strong labor protection laws. However, since much of the growth in the developing world has been inequitable, the World Bank has also begun teaming with client states to map out trends in inequality and to propose public policy changes that can level the playing field. Moreover, the World Bank engages in a variety of nutritional, transfer payments and transport-based initiatives. Children who experience under-nutrition from conception to two years of age have a much higher risk of physical and mental disability. Thus, they are often trapped in poverty and are unable to make a full contribution to the social and economic development of their communities as adults. The WBG estimates that as much as 3% of GDP can be lost as a result of under-nutrition among the poorest nations. To combat undernutrition, the WBG has partnered with UNICEF and the WHO to ensure all small children are fully fed. The WBG also offers conditional cash transfers to poor households who meet certain requirements such as maintaining children's healthcare or ensuring school attendance. Finally, the WBG understands investment in public transportation and better roads is key to breaking rural isolation, improving access to healthcare and providing better job opportunities for the World's poor. United Nations The UN Office for the Coordination of Humanitarian Affairs (OCHA) works to synchronize the disparate international, national and non-governmental efforts to contest poverty. OCHA seeks to prevent "confusion" in relief operations and to ensure that the humanitarian response to disaster situations has greater accountability and predictability. To do so, OCHA has begun deploying Humanitarian Coordinators and Country Teams to provide a solid architecture for the international community to work through. The United Nation's Children's Fund (UNICEF) was created by the UN to provide food, clothing and healthcare to European children facing famine and disease in the immediate aftermath of World War II. After the UN General Assembly extended UNICEF's mandate indefinitely in 1953, it actively worked to help children in extreme poverty in more than 190 countries and territories to overcome the obstacles that poverty, violence, disease and discrimination place in a child's path. Its current focus areas are 1) Child survival & development 2) Basic education & gender equality 3) Children and HIV/AIDS and 4) Child protection. The UN Refugee Agency (UNHCR) is mandated to lead and coordinate international action to protect refugees worldwide. Its primary purpose is to safeguard the rights of refugees by ensuring anyone can exercise the right to seek asylum in another state, with the option to return home voluntarily, integrate locally or resettle in a third country. The UNHCR operates in over 125 countries, helping approximately 33.9 million persons. The World Food Programme (WFP) is the largest agency dedicated to fighting hunger worldwide. On average, the WFP brings food assistance to more than 90 million people in 75 countries. The WFP not only strives to prevent hunger in the present, but also in the future by developing stronger communities which will make food even more secure on their own. The WFP has a range of expertise from Food Security Analysis, Nutrition, Food Procurement and Logistics. The World Health Organization (WHO) is responsible for providing leadership on global health matters, shaping the health research agenda, articulating evidence-based policy decisions and combating diseases that are induced from poverty, such as HIV/AIDS, malaria and tuberculosis. Moreover, the WHO deals with pressing issues ranging from managing water safety, to dealing with maternal and newborn health. Governmental agencies USAID The US Agency for International Development (USAID) is the lead US government agency dedicated to ending extreme poverty. Currently the largest bilateral donor in the world, the United States channels the majority of its development assistance through USAID and the US Department of State. In President Obama's 2013 State of the Union address, he declared, "So the United States will join with our allies to eradicate such extreme poverty in the next two decades ... which is within our reach." In response to Obama's call to action, USAID has made ending extreme poverty central to its mission statement. Under its New Model of Development, USAID seeks to eradicate extreme poverty through the use of innovation in science and technology, by putting a greater emphasis on evidence based decision-making, and through leveraging the ingenuity of the private sector and global citizens. A major initiative of the Obama Administration is Power Africa, which aims to bring energy to 20 million people in Sub-Saharan Africa. By reaching out to its international partners, whether commercial or public, the US has leveraged over $14 billion in outside commitments after investing only US$7 billion of its own. To ensure that Power Africa reaches the region's poorest, the initiative engages in a transaction based approach to create systematic change. This includes expanding access to electricity to more than 20,000 additional households which already live without power. In terms of specific programming, USAID works in a variety of fields from preventing hunger, reducing HIV/AIDS, providing general health assistance and democracy assistance, as well as dealing with gender issues. To deal with food security, which affects roughly 842 million people (who go to bed hungry each night), USAID coordinates the Feed the Future Initiative (FtF). FtF aims to reduce poverty and under-nutrition each by 20% over five years. Because of the President's Emergency Plan for AIDS Relief (PEPFAR) and a variety of congruent actors, the incidence of AIDS and HIV, which used to ravage Africa, reduced in scope and intensity. Through PEPFAR, the United States has ensured over five million people have received life-saving antiviral drugs, a significant proportion of the eight million people receiving treatment in relatively poor nations. In terms of general health assistance, USAID has worked to reduce maternal mortality by 30%, under-five child mortality by 35%, and has accomplished a host of other goals. USAID also supports the gamut of democratic initiatives, from promoting human rights and accountable, fair governance, to supporting free and fair elections and the rule of law. In pursuit of these goals, USAID has increased global political participation by training more than 9,800 domestic election observers and providing civic education to more than 6.5 million people. Since 2012, the Agency has begun integrating critical gender perspectives across all aspects of its programming to ensure all USAID initiatives work to eliminate gender disparities. To do so, USAID seeks to increase the capability of women and girls to realize their rights and determine their own life outcomes. Moreover, USAID supports additional programs to improve women's access to capital and markets, builds theirs skills in agriculture, and supports women's desire to own businesses. Others Other major government development agencies with annual aid programmes of more than $10 billion include: GIZ (Germany), FCDO (United Kingdom), JICA (Japan), European Union and AFD (France). Non-Governmental Organizations A multitude of non-governmental organizations operate in the field of extreme poverty, actively working to alleviate the poorest of the poor of their deprivation. To name but a few notable organizations: Save the Children, the Overseas Development Institute, Concern Worldwide, ONE, Trickle Up and Oxfam have all done a considerable amount of work in extreme poverty. Save the Children is the leading international organization dedicated to helping the world's indigent children. In 2013, Save the Children reached over 143 million children through their work, including over 52 million children directly. Save the Children also recently released their own report titled "Getting to Zero", in which they argued the international community could feasibly do more than lift the world's poor above $1.25/day. The Overseas Development Institute (ODI) is a UK based think tank on international development and humanitarian issues. ODI is dedicated to alleviating the suffering of the world's poor by providing high-quality research and practical policy advice to the World's development officials. ODI also recently released a paper entitled, "The Chronic Poverty Report 2014–2015: The road to zero extreme poverty", in which its authors assert that though the international communities' goal of ending extreme poverty by 2030 is laudable, much more targeted resources will be necessary to reach said target. The report states that "To eradicate extreme poverty, massive global investment is required in social assistance, education and pro-poorest economic growth". Concern Worldwide is an international humanitarian organization whose mission is to end extreme poverty by influencing decision makers at all levels of government (from local to international). Concern has also produced a report on extreme poverty in which they explain their own conception of extreme poverty from a NGO's standpoint. In this paper, named "How Concern Understands Extreme Poverty", the report's creators write that extreme poverty entails more than just living under $1.25/day, it also includes having a small number of assets and being vulnerable to severe negative shocks (whether natural or man made). ONE, the organization co-founded by Bono, is a non-profit organization funded almost entirely by foundations, individual philanthropists and corporations. ONE's goals include raising public awareness and working with political leaders to fight preventable diseases, increase government accountability and increase investment in nutrition. Finally, trickleUp is a micro-enterprise development program targeted at those living on under $1.25/day, which provides the indigent with resources to build a sustainable livelihood through both direct financing and considerable training efforts. Oxfam is a non-governmental organization that works prominently in Africa; their mission is to improve local community organizations and it works to reduce impediments to the development of the country. Oxfam helps families suffering from poverty receive food and healthcare to survive. There are many children in Africa experiencing growth stunting, and this is one example of an issue that Oxfam targets and aims to resolve. Cash transfers appear to be an effective intervention for reducing extreme poverty, while at the same time improving health and education outcomes. Campaigns Giving What We Can Global Poverty Project Live Below the Line Make Poverty History See also List of countries by percentage of population living in poverty Income inequality metrics Least developed countries Poverty threshold Poverty reduction Millennium Development Goals (2015) Sustainable Development Goals (2030) Notes References External links Eradicate Extreme Poverty and Hunger by 2015 | UN Millennium Development Goal curated by the Center for Latin American and Caribbean Studies at Michigan State University The Life You Can Save – Acting Now to End World Poverty Scientific American Magazine (September 2005 Issue) Can Extreme Poverty Be Eliminated? International Movement ATD Fourth World Walk In Her Shoes Measurements and definitions of poverty
1435
https://en.wikipedia.org/wiki/Abbotsford%2C%20Scottish%20Borders
Abbotsford, Scottish Borders
Abbotsford is a historic country house in the Scottish Borders, near Galashiels, on the south bank of the River Tweed. Now open to the public, it was built as the residence of historical novelist and poet Sir Walter Scott between 1817 and 1825. It is a Category A Listed Building and the estate is listed in the Inventory of Gardens and Designed Landscapes in Scotland. Description The nucleus of the estate was a small farm of , called Cartleyhole, nicknamed Clarty (i.e., muddy) Hole, and was bought by Scott on the lapse of his lease (1811) of the neighbouring house of Ashestiel. Scott renamed it "Abbotsford" after a neighbouring ford used by the monks of Melrose Abbey. Following a modest enlargement of the original farmhouse in 1811–12, massive expansions took place in 1816–19 and 1822–24. In this mansion Scott gathered a large library, a collection of ancient furniture, arms and armour, and other relics and curiosities especially connected with Scottish history, notably the Celtic Torrs Pony-cap and Horns and the Woodwrae Stone, all now in the Museum of Scotland. Scott described the resulting building as "a sort of romance in Architecture" and "a kind of Conundrum Castle to be sure". The last and principal acquisition was that of Toftfield (afterwards named Huntlyburn), purchased in 1817. The new house was then begun and completed in 1824. The general ground-plan is a parallelogram, with irregular outlines, one side overlooking the Tweed; and the style is mainly the Scottish Baronial. With his architects William Atkinson and Edward Blore Scott was a pioneer of the Scottish Baronial style of architecture: the house is recognized as a highly influential creation with themes from Abbotsford being reflected across many buildings in the Scottish Borders and beyond. The manor as a whole appears as a "castle-in-miniature", with small towers and imitation battlements decorating the house and garden walls. Into various parts of the fabric were built relics and curiosities from historical structures, such as the doorway of the old Tolbooth in Edinburgh. Scott collected many of these curiosities to be built into the walls of the South Garden, which previously hosted a colonnade of gothic arches along the garden walls. Along the path of the former colonnade sits the remains of Edinburgh's 15th century Mercat Cross and several examples of classical sculpture. The estate and its neo-Medieval features nod towards Scott's desire for a historical feel, but the writer ensured that the house would provide all the comforts of modern living. As a result, Scott used the space as a proving-ground for new technologies. The house was outfitted with early gas lighting and pneumatic bells connecting residents with servants elsewhere in the house. Scott had only enjoyed his residence one year when (1825) he met with that reverse of fortune which involved the estate in debt. In 1830, the library and museum were presented to him as a free gift by the creditors. The property was wholly disencumbered in 1847 by Robert Cadell, the publisher, who cancelled the bond upon it in exchange for the family's share in the copyright of Sir Walter's works. Scott's only son Walter did not live to enjoy the property, having died on his way from India in 1847. Among subsequent possessors were Scott's grandson Walter Scott Lockhart (later Walter Lockhart Scott, 1826–1853), his younger sister Charlotte Harriet Jane Hope-Scott (née Lockhart) 1828–1858, J. R. Hope Scott, QC, and his daughter (Scott's great-granddaughter), the Hon. Mrs Maxwell Scott. The house was opened to the public in 1833, but continued to be occupied by Scott's descendants until 2004. The last of his direct descendants to hold the Lairdship of Abbotsford was his great-great-great-granddaughter Dame Jean Maxwell-Scott (8 June 1923 – 5 May 2004). She inherited it from her elder sister Patricia Maxwell-Scott in 1998. The sisters turned the house into one of Scotland's premier tourist attractions, after they had to rely on paying visitors to afford the upkeep of the house. It had electricity installed only in 1962. Dame Jean was at one time a lady-in-waiting to Princess Alice, Duchess of Gloucester, patron of the Dandie Dinmont Club, a breed of dog named after one of Sir Walter Scott's characters; and a horse trainer, one of whose horses, Sir Wattie, ridden by Ian Stark, won two silver medals at the 1988 Summer Olympics. On Dame Jean's death the Abbotsford Trust was established to safeguard the estate. In 2005, Scottish Borders Council considered an application by a property developer to build a housing estate on the opposite bank of the River Tweed from Abbotsford, to which Historic Scotland and the National Trust for Scotland objected. There have been modifications to the proposed development, but it is still being opposed in 2020. Sir Walter Scott rescued the "jougs" from Threave Castle in Dumfries and Galloway and attached them to the castellated gateway he built at Abbotsford. Tweedbank railway station is located near to Abbotsford. Miscellaneous Abbotsford gave its name to the Abbotsford Club, founded by William Barclay Turnbull in 1833 or 1834 in Scott's honour, and a successor to the Bannatyne and Maitland Clubs. It was a text publication society, which existed to print and publish historical works connected with Scott's writings. Its publications extended from 1835 to 1864. In August 2012, a new Visitor Centre opened at Abbotsford which houses a small exhibition, gift shop and Ochiltree's café with views over the house and grounds. The house re-opened to the public after extensive renovations in July 2013. In 2014 it won the European Union Prize for Cultural Heritage / Europa Nostra Award for its recent conservation project. See also List of places in the Scottish Borders Notes References Attribution External links Abbotsford – The Home of Sir Walter Scott – official site RCAHMS / CANMORE site record for Abbotsford Edinburgh University Library Abbotsford (by W S Crockett – 1904 illustrated book pub. A & C Black) Abbotsford and Newstead Abbey by Washington Irving, from Project Gutenberg Category A listed buildings in the Scottish Borders Category A listed houses in Scotland Scottish baronial architecture Walter Scott Inventory of Gardens and Designed Landscapes Gardens in the Scottish Borders Literary museums in Scotland Historic house museums in the Scottish Borders Country houses in the Scottish Borders Houses completed in 1824 Galashiels
1576
https://en.wikipedia.org/wiki/Afonso%20de%20Albuquerque
Afonso de Albuquerque
Afonso de Albuquerque, 1st Duke of Goa ( – 16 December 1515), was a Portuguese general, admiral, and statesman. He served as viceroy of Portuguese India from 1509 to 1515, during which he expanded Portuguese influence across the Indian Ocean and built a reputation as a fierce and skilled military commander. Albuquerque advanced the three-fold Portuguese grand scheme of combating Islam, spreading Christianity, and securing the trade of spices by establishing a Portuguese Asian empire. Among his achievements, Albuquerque managed to conquer Goa and was the first European of the Renaissance to raid the Persian Gulf, and he led the first voyage by a European fleet into the Red Sea. He is generally considered a highly effective military commander, and "probably the greatest naval commander of the age", given his successful strategy — he attempted to close all the Indian Ocean naval passages to the Atlantic, Red Sea, Persian Gulf, and to the Pacific, transforming it into a Portuguese mare clausum. He was appointed head of the "fleet of the Arabian and Persian sea" in 1506. Many of the conflicts in which he was directly involved took place in the Indian Ocean, in the Persian Gulf regions for control of the trade routes, and on the coasts of India. His military brilliance in these initial campaigns enabled Portugal to become the first global empire in history. He led the Portuguese forces in numerous battles, including the conquest of Goa in 1510 and the capture of Malacca in 1511. During the last five years of his life, he turned to administration, where his actions as the second governor of Portuguese India were crucial to the longevity of the Portuguese Empire. He oversaw expeditions that resulted in establishing diplomatic contacts with the Ayutthaya Kingdom through his envoy Duarte Fernandes, with Pegu in Myanmar, and Timor and the Moluccas through a voyage headed by António de Abreu and Francisco Serrão. He laid the path for European trade with Ming China through Rafael Perestrello. He also aided in establishing diplomatic relations with Ethiopia, and established diplomatic ties with Persia during the Safavid dynasty. Throughout his career, he received epithets such as "the Terrible", "the Great", "the Lion of the Seas", "the Portuguese Mars", and "the Caesar of the East". Early life Afonso de Albuquerque was born in 1453 in Alhandra, near Lisbon. He was the second son of Gonçalo de Albuquerque, Lord of Vila Verde dos Francos, and Dona Leonor de Menezes. His father held an important position at court and was connected by remote illegitimate descent with the Portuguese monarchy. He was a descendant of King Denis’s illegitimate son, Afonso Sanches, Lord of Albuquerque. He was educated in mathematics and Latin at the court of Afonso V of Portugal, where he befriended Prince John, the future King John II of Portugal. Early military service In 1471, under the command of Afonso V, he was present at the conquest of Tangier and Arzila in Morocco, and he served there as an officer for some years. In 1476, he accompanied Prince John in wars against Castile, including the Battle of Toro. He participated in the campaign on the Italian peninsula in 1480 to assist Ferdinand I of Naples in repelling the Ottoman invasion of Otranto. On his return in 1481, when John was crowned as King John II, Albuquerque was made master of the horse and chief equerry () to the king, a post which he held throughout John's reign. In 1489, he returned to military campaigning in North Africa, as commander of defense in the Graciosa fortress, an island in the river Luco near the city of Larache. In 1490 Albuquerque was part of the guard of John II. He returned to Arzila in 1495, where his younger brother Martim died fighting by his side. First expedition to India, 1503 When King Manuel I of Portugal ascended to the throne following the death of his cousin John II, he held a cautious attitude towards Albuquerque, who was a close friend of his predecessor and seventeen years Manuel's senior. Eight years later, on 6 April 1503 Albuquerque was sent on his first expedition to India together with his cousin Francisco de Albuquerque. Each commanded three ships, sailing with Duarte Pacheco Pereira and Nicolau Coelho. They engaged in several battles against the forces of the Zamorin of Calicut (Calecute, Kozhikode) and succeeded in establishing the king of Cochin (Cohim, Kochi) securely on his throne. In return, the king of Cochin gave the Portuguese permission to build the Portuguese fort Immanuel (Fort Kochi) and establish trade relations with Quilon (Coulão, Kollam). This laid the foundation for the eastern Portuguese Empire. Second expedition to India, 1506 Albuquerque returned home in July 1504 and was well received by King Manuel I. After he assisted with the creation of a strategy for the Portuguese efforts in the east, King Manuel entrusted him with the command of a squadron of five vessels in the fleet of sixteen sailing for India in early 1506, headed by Tristão da Cunha. The aim of the expedition was to conquer Socotra and build a fortress there, hoping to close the trade in the Red Sea. Albuquerque went as "chief-captain for the Coast of Arabia", sailing under da Cunha's orders until reaching Mozambique. He carried a sealed letter with a secret mission ordered by the king: after fulfilling the first mission, he was to replace the first viceroy of India, Francisco de Almeida, whose term ended two years later. Before departing, he legitimized his son Brás ("Braz" in the old Portuguese spelling), born to a common Portuguese woman named Joana Vicente in 1500. First conquest of Socotra and Ormuz, 1507 The fleet left Lisbon on 6 April 1506. Albuquerque piloted his ship himself, having lost his appointed pilot on departure. In Mozambique Channel, they rescued Captain João da Nova, who had encountered difficulties on his return from India; da Nova and his ship, the Frol de la mar, joined da Cunha's fleet. From Malindi, da Cunha sent envoys to Ethiopia, which at the time was thought to be closer to India than it actually is, under the aegis of Albuquerque. After failing to reach Ethiopia, he managed to land the envoys in Filuk. After successful attacks on Arab cities on the East African coast, the expedition conquered the island of Socotra and built a fortress at Suq, hoping to establish a base to stop the Red Sea commerce to the Indian Ocean. However, Socotra was abandoned four years later, as it was eventually realised to be a poor location for a base. At Socotra, they parted ways: Tristão da Cunha sailed for India, where he would relieve the Portuguese besieged at Cannanore, while Afonso took seven ships and 500 men to Ormuz in the Persian Gulf, one of the chief eastern centers of commerce. On his way, he conquered the cities of Curiati (Kuryat), Muscat in July 1507, and Khor Fakkan, accepting the submission of the cities of Kalhat and Sohar. He arrived at Hormuz on 25 September and soon captured the city, which agreed to become a tributary state of the Portuguese king. Ormuz was then a tributary state of Shah Ismail of Persia. In a famous episode, shortly after its conquest, Albuquerque was confronted by Persian envoys, who demanded the payment of the due tribute from him instead. He ordered them to be given a stock of cannonballs, arrows and weapons, retorting that "such was the currency struck in Portugal to pay the tribute demanded from the dominions of King Manuel". According to Brás de Albuquerque, it was Shah Ismael who first addressed Albuquerque as "Lion of the seas". Afonso began building the Fort of Our Lady of Victory (later renamed Fort of Our Lady of the Conception) on Hormuz Island, engaging his men of all ranks in the work. However, some of his officers, claiming that Afonso was exceeding his orders, revolted against the heavy work and climate and departed for India. With his fleet reduced to two ships and left without supplies, he was unable to maintain his position. In January 1508, he was forced to abandon Ormuz. He raided coastal villages to resupply the settlement of Socotra, returned to Ormuz, and then headed to India. Arrest at Cannanore, 1509 Afonso arrived at Cannanore on the Malabar coast in December 1508, where he opened the sealed letter that he had received from the king before the viceroy, Dom Francisco de Almeida, which named him as governor to succeed Almeida. The viceroy, supported by the officers who had abandoned Afonso at Ormuz, had a matching royal order but declined to yield. He protested that his term ended only in January and stated his intention to avenge his son's death by fighting the Mamluk fleet of Mirocem, refusing Afonso's offer to fight the Mamluk fleet himself. Afonso avoided confrontation, which could have led to civil war, and moved to Kochi, India, to await further instruction from the king. Increasingly isolated, he wrote to Diogo Lopes de Sequeira, who arrived in India with a new fleet, but was ignored as Sequeira joined Almeida. At the same time, Afonso refused approaches from opponents of Almeida who encouraged him to seize power. On 3 February 1509, Almeida fought the naval Battle of Diu against a joint fleet of Mamluks, Ottomans, the Zamorin of Calicut, and the Sultan of Gujarat. His victory was decisive: the Ottomans and Mamluks abandoned the Indian Ocean, easing the way for Portuguese rule there for the next century. In August, after a petition from Afonso's former officers with the support of Diogo Lopes de Sequeira claiming him unfit for governance, Afonso was sent in custody to St. Angelo Fort in Cannanore. There he remained under what he considered as imprisonment. Governor of Portuguese India, 1509–1515 Afonso was released after three months' confinement, on the arrival at Cannanore of the Marshal of Portugal Fernando Coutinho with a large fleet sent by the king. Coutinho was the most important Portuguese noble to visit India up to that point. He brought an armada of fifteen ships and 3,000 men to defend Afonso's rights, and to take Calicut. On 4 November 1509, Afonso became the second Governor of Portuguese India, a position he would hold until his death. Almeida set off to return to Portugal, but he was killed before he got there in a skirmish with the Khoekhoe. Upon his assuming office, Afonso intended to dominate the Muslim world and control the Spice trade. Initially, King Manuel I and his council in Lisbon tried to distribute the power by outlining three areas of jurisdiction in the Indian Ocean. In 1509, the nobleman Diogo Lopes de Sequeira was sent with a fleet to Southeast Asia, to seek an agreement with Sultan Mahmud Shah of Malacca, but failed and returned to Portugal. To Jorge de Aguiar was given the region between the Cape of Good Hope and Gujarat. He was succeeded by Duarte de Lemos, but left for Cochin and then for Portugal, leaving his fleet to Afonso. Conquest of Goa, 1510 In January 1510, obeying the orders from the king and aware of the absence of the Zamorin, Afonso advanced on Calicut. The attack was initially successful, but unravelled when Marshal Coutinho, infuriated by Albuquerque's success against Calicut and desiring glory for himself, attacked the Zamorin's palace against Albuquerque's advice, and was ambushed. During the retreat, Afonso was badly wounded and was forced to flee to the ships, barely escaping with his life, while Coutinho was killed. Soon after the failed attack, Afonso assembled a fleet of 23 ships and 1200 men. Contemporary reports state that he wanted to fight the Egyptian Mamluk Sultanate fleet in the Red Sea or return to Hormuz. However, he had been informed by Timoji (a privateer in the service of the Hindu Vijayanagara Empire) that it would be easier to fight them in Goa, where they had sheltered after the Battle of Diu, and also of the illness of the Sultan Yusuf Adil Shah, and war between the Deccan sultanates. So he relied on surprise in the capture of Goa from the Sultanate of Bijapur. A first assault took place in Goa from 4 March to 20 May 1510. After the initial occupation, feeling unable to hold the city given the poor condition of its fortifications, the cooling of Hindu residents' support and insubordination among his ranks following an attack by Ismail Adil Shah, Afonso refused a truce offered by the Sultan and abandoned the city in August. His fleet was scattered, and a palace revolt in Kochi hindered his recovery, so he headed to Fort Anjediva. New ships arrived from Portugal, which were intended for the nobleman Diogo Mendes de Vasconcelos at Malacca, who had been given a rival command of the region. Three months later, on 25 November Afonso reappeared at Goa with a renovated fleet. Diogo Mendes de Vasconcelos was compelled to accompany him with the reinforcements for Malacca and about 300 Malabari reinforcements from Cannanore. In less than a day, they took Goa from Ismail Adil Shah and his Ottoman allies, who surrendered on 10 December. It is estimated that 6000 of the 9000 Muslim defenders of the city died, either in the fierce battle in the streets or by drowning while trying to escape. Afonso regained the support of the Hindu population, although he frustrated the initial expectations of Timoji, who aspired to become governor. Afonso rewarded him by appointing him chief "Aguazil" of the city, an administrator and representative of the Hindu and Muslim people, as a knowledgeable interpreter of the local customs. He then made an agreement to lower the yearly tribute. In Goa, Afonso established the first Portuguese mint in the East, after Timoja's merchants had complained of the scarcity of currency, taking it as an opportunity to solidify the territorial conquest. The new coin, based on the existing local coins, showed a cross on the obverse and an armillary sphere (or "esfera"), King Manuel's badge, on the reverse. Gold cruzados or manueis, silver esferas and alf-esferas, and bronze "leais" were issued. Albuquerque founded at Goa the Hospital Real de Goa or Royal Hospital of Goa, by the Church of Santa Catarina. Upon hearing that the doctors were extorting the sickly with excessive fees, Albuquerque summoned them, declaring that "You charge a physician's pay and don't know what disease the men who serve our lord the King suffer from. Thus, I want to teach you what is it that they die from" and put them to work building the city walls all day till nightfall before releasing them. Despite constant attacks, Goa became the center of Portuguese India, with the conquest triggering the compliance of neighbouring kingdoms: the Sultan of Gujarat and the Zamorin of Calicut sent embassies, offering alliances and local grants to fortify. Afonso then used Goa to secure the spice trade in favor of Portugal and sell Persian horses to Vijayanagara and Hindu princes in return for their assistance. Conquest of Malacca, 1511 Afonso explained to his armies why the Portuguese wanted to capture Malacca: "The King of Portugal has often commanded me to go to the Straits, because...this was the best place to intercept the trade which the Moslems...carry on in these parts. So it was to do Our Lord's service that we were brought here; by taking Malacca, we would close the Straits so that never again would the Moslems be able to bring their spices by this route.... I am very sure that, if this Malacca trade is taken out of their hands, Cairo and Mecca will be completely lost." (The Commentaries of the Great Afonso de Albuquerque) In February 1511, through a friendly Hindu merchant, Nina Chatu, Afonso received a letter from Rui de Araújo, one of the nineteen Portuguese held at Malacca since 1509. It urged moving forward with the largest possible fleet to demand their release, and gave details of the fortifications. Afonso showed it to Diogo Mendes de Vasconcelos, as an argument to advance as a joint fleet. In April 1511, after fortifying Goa, he gathered a force of about 900 Portuguese, 200 Hindu mercenaries and about eighteen ships. He then sailed to Malacca against orders and despite the protest of Diogo Mendes, who claimed command of the expedition. Afonso eventually centralized the Portuguese government in the Indian Ocean. After the Malaccan conquest, he wrote a letter to the king to explain his disagreement with Diogo Mendes, suggesting that further divisions could be harmful to the Portuguese in India. Under his command was Ferdinand Magellan, who had participated in the failed embassy of Diogo Lopes de Sequeira in 1509. After a false start towards the Red Sea, they sailed to the Strait of Malacca. It was the richest city that the Portuguese tried to take, and a focal point in the trade network where Malay traders met Gujarati, Chinese, Japanese, Javanese, Bengali, Persian and Arabic, among others, described by Tomé Pires as of invaluable richness. Despite its wealth, it was mostly a wooden-built city, with few masonry buildings but was defended by a mercenary force estimated at 20,000 men and more than 2000 pieces of artillery. Its greatest weakness was the unpopularity of the government of Sultan Mahmud Shah, who favoured Muslims, arousing dissatisfaction amongst other merchants. Afonso made a bold approach to the city, his ships decorated with banners, firing cannon volleys. He declared himself lord of all the navigation, demanded the Sultan release the prisoners and pay for damages, and demanded consent to build a fortified trading post. The Sultan eventually freed the prisoners, but was unimpressed by the small Portuguese contingent. Afonso then burned some ships at the port and four coastal buildings as a demonstration. The city being divided by the Malacca River, the connecting bridge was a strategic point, so at dawn on 25 July, the Portuguese landed and fought a tough battle, facing poisoned arrows, taking the bridge in the evening. After fruitlessly waiting for the Sultan's reaction, they returned to the ships and prepared a junk (offered by Chinese merchants), filling it with men, artillery and sandbags. Commanded by António de Abreu, it sailed upriver at high tide to the bridge. The day after, all had landed. After a fierce fight during which the Sultan appeared with an army of war elephants, the defenders were dispersed and the Sultan fled. Afonso waited for the reaction of the Sultan. Merchants approached, asking for Portuguese protection. They were given banners to mark their premises, a sign that they would not be looted. On 15 August, the Portuguese attacked again, but the Sultan had fled the city. Under strict orders, they looted the city, but respected the banners. Afonso prepared Malacca's defenses against a Malay counterattack, building a fortress, assigning his men to shifts and using stones from the mosque and the cemetery. Despite the delays caused by heat and malaria, it was completed in November 1511, its surviving door now known as "A Famosa" ('the famous'). It was possibly then that Afonso had a large stone engraved with the names of the participants in the conquest. To quell disagreements over the order of the names, he had it set facing the wall, with the single inscription Lapidem quem reprobaverunt aedificantes (Latin for "The stone the builders rejected", from David's prophecy, Psalm 118:22–23) on the front. He settled the Portuguese administration, reappointing Rui de Araújo as factor, a post assigned before his 1509 arrest, and appointing rich merchant Nina Chatu to replace the previous Bendahara. Besides assisting in the governance of the city and the first Portuguese coinage, he provided the junks for several diplomatic missions. Meanwhile, Afonso arrested and had executed the powerful Javanese merchant Utimuti Raja who, after being appointed to a position in the Portuguese administration as representative of the Javanese population, had maintained contacts with the exiled royal family. Shipwreck on the Flor de la mar, 1511 On 20 November 1511 Afonso sailed from Malacca to the coast of Malabar on the old Flor de la Mar carrack that had served to support the conquest of Malacca. Despite its unsound condition, he used it to transport the treasure amassed in the conquest, given its large capacity. He wanted to give the court of King Manuel a show of Malaccan treasures. There were also offerings from the Ayutthaya Kingdom (Thailand) to the king of Portugal, and all his own fortune. On the voyage, the Flor de la Mar was wrecked in a storm, and Afonso barely escaped drowning. Missions from Malacca Embassies to Pegu, Sumatra and Siam, 1511 Most Muslim and Gujarati merchants having fled the city, Afonso invested in diplomatic efforts demonstrating generosity to Southeast Asian merchants, like the Chinese, to encourage good relations with the Portuguese. Trade and diplomatic missions were sent to continental kingdoms: Rui Nunes da Cunha was sent to Pegu (Burma), from where King Binyaram sent back a friendly emissary to Kochi in 1514 and Sumatra, Sumatran kings of Kampar and Indragiri sending emissaries to Afonso accepting the new power, as vassal states of Malacca. Knowing of Siamese ambitions over Malacca, Afonso sent Duarte Fernandes in a diplomatic mission to the Ayutthaya Kingdom (Thailand), returning in a Chinese junk. He was one of the Portuguese who had been arrested in Malacca, having gathered knowledge about the culture of the region. There he was the first European to arrive, establishing amicable relations between the kingdom of Portugal and the court of the king of Siam Ramathibodi II, returning with a Siamese envoy bearing gifts and letters to Afonso and the king of Portugal. Expedition to the "spice islands" (Maluku islands), 1512 In November, after having secured Malacca and learning the location of the then secret "spice islands", Afonso sent three ships to find them, led by trusted António de Abreu with deputy commander Francisco Serrão. Malay sailors were recruited to guide them through Java, the Lesser Sunda Islands and the Ambon Island to Banda Islands, where they arrived in early 1512. There they remained for a month, buying and filling their ships with nutmeg and cloves. António de Abreu then sailed to Amboina whilst Serrão sailed towards the Moluccas, but he was shipwrecked near Seram. Sultan Abu Lais of Ternate heard of their stranding, and, seeing a chance to ally himself with a powerful foreign nation, brought them to Ternate in 1512 where they were permitted to build a fort on the island, the , built in 1522. Return to Cochin and Goa Afonso returned from Malacca to Cochin, but could not sail to Goa as it faced a serious revolt headed by the forces of Ismael Adil Shah, the Sultan of Bijapur, commanded by Rasul Khan and his countrymen. During Afonso's absence from Malacca, the Portuguese who opposed the taking of Goa had waived its possession, even writing to the king that it would be best to let it go. Held up by the monsoon and with few forces available, Afonso had to wait for the arrival of reinforcement fleets headed by his nephew D. Garcia de Noronha, and Jorge de Mello Pereira. While at Cochin, Albuquerque started a school. In a private letter to King Manuel I, he stated that he had found a chest full of books with which to teach the children of married Portuguese settlers (casados) and Christian converts, of which there were about a hundred, to read and write. On 10 September 1512, Afonso sailed from Cochin to Goa with fourteen ships carrying 1,700 soldiers. Determined to recapture the fortress, he ordered trenches dug and a wall breached. But on the day of the planned final assault, Rasul Khan surrendered. Afonso demanded the fort be handed over with its artillery, ammunition and horses, and the deserters to be given up. Some had joined Rasul Khan when the Portuguese were forced to flee Goa in May 1510, others during the recent siege. Rasul Khan consented, on condition that their lives be spared. Afonso agreed and he left Goa. He did spare the lives of the deserters, but had them horribly mutilated. One such renegade was Fernão Lopes, bound for Portugal in custody, who escaped at the island of Saint Helena and led a 'Robinson Crusoe' life for many years. After such measures the town became the most prosperous Portuguese settlement in India. Campaign in the Red Sea, 1513 In December 1512 an envoy from Ethiopia arrived at Goa. Mateus was sent by the regent queen Eleni, following the arrival of the Portuguese from Socotra in 1507, as an ambassador for the king of Portugal in search of a coalition to help face growing Muslim influence. He was received in Goa with great honour by Afonso, as a long-sought "Prester John" envoy. His arrival was announced by King Manuel to Pope Leo X in 1513. Although Mateus faced the distrust of Afonso's rivals, who tried to prove he was some impostor or Muslim spy, Afonso sent him to Portugal. The king is described as having wept with joy at their report. In February 1513, while Mateus was in Portugal, Afonso sailed to the Red Sea with a force of about 1000 Portuguese and 400 Malabaris. He was under orders to secure that channel for Portugal. Socotra had proved ineffective to control the Red Sea entrance and was abandoned, and Afonso's hint that Massawa could be a good Portuguese base might have been influenced by Mateus' reports. Knowing that the Mamluks were preparing a second fleet at Suez, he wanted to advance before reinforcements arrived in Aden, and accordingly laid siege to the city. Aden was a fortified city, but although he had scaling ladders they broke during the chaotic attack. After half a day of fierce battle, Afonso was forced to retreat. He cruised the Red Sea inside the Bab al-Mandab, with the first European fleet to have sailed this route. He attempted to reach Jeddah, but the winds were unfavourable and so he sheltered at Kamaran island in May, until sickness among the men and lack of fresh water forced him to retreat. In August 1513, after a second attempt to reach Aden, he returned to India with no substantial results. In order to destroy the power of Egypt, he wrote to King Manuel of the idea of diverting the course of the Nile river to render the whole country barren. He also intended to steal the body of the Islamic prophet, Muhammad, and hold it for ransom until all Muslims had left the Holy Land. Although Albuquerque's expedition failed to reach Suez, such an incursion into the Red Sea by a Christian fleet for the first time in history stunned the Muslim world, and panic spread in Cairo. Submission of Calicut Albuquerque achieved during his term a favourable end to hostilities between the Portuguese and the Zamorin of Calicut, which had lasted since the massacre of the Portuguese in Calicut in 1502. As naval trade faltered and vassals defected, with no foreseeable solutions to the conflict with the Portuguese, the court of the Zamorin fell to in-fighting. The ruling Zamorin was assassinated and replaced by a rival, under the instigation of Albuquerque, permitting peace talks to commence. The Portuguese were allowed to build a fortress in Calicut itself, and acquired rights to obtain as much pepper and ginger as they wished, at stipulated prices, and half the customs duties of Calicut as yearly tribute. Construction of the fortress began immediately, under the supervision of chief architect Tomás Fernandes. Administration and diplomacy in Goa, 1514 With peace concluded, in 1514 Afonso devoted himself to governing Goa and receiving embassies from Indian governors, strengthening the city and encouraging marriages of Portuguese men and local women. At that time, Portuguese women were barred from traveling overseas in order to maintain discipline among the men on board the ships. In 1511 under a policy which Afonso promulgated, the Portuguese government encouraged their explorers to marry local women. To promote settlement, the King of Portugal granted freeman status and exemption from Crown taxes to Portuguese men (known as casados, or "married men") who ventured overseas and married local women. With Afonso's encouragement, mixed marriages flourished, giving birth to Portuguese-Indians or mestiços. He appointed local people for positions in the Portuguese administration and did not interfere with local traditions (except "sati", the practice of immolating widows, which he banned). In March 1514 King Manuel sent to Pope Leo X a huge and exotic embassy led by Tristão da Cunha, who toured the streets of Rome in an extravagant procession of animals from the colonies and wealth from the Indies. His reputation reached its peak, laying foundations of the Portuguese Empire in the East. In early 1514, Afonso sent ambassadors to Gujarat's Sultan Muzaffar Shah II, ruler of Cambay, to seek permission to build a fort on Diu, India. The mission returned without an agreement, but diplomatic gifts were exchanged, including an Indian rhinoceros, Afonso sent the rhino to King Manuel, making it the first living example of a rhinoceros seen in Europe since the Roman Empire. Conquest of Ormuz and Illness In 1513, at Cannanore, Afonso was visited by a Persian ambassador from Shah Ismail I, who had sent ambassadors to Gujarat, Ormuz and Bijapur. The shah's ambassador to Bijapur invited Afonso to send back an envoy to Persia. Miguel Ferreira was sent via Ormuz to Tabriz, where he had several interviews with the shah about common goals of defeating the Mamluk sultan. At the same time, Albuquerque decided to conclude the effective conquest of Hormuz. He had learned that after the Portuguese retreat in 1507, a young king was reigning under the influence of a powerful Persian vizier, Reis Hamed, whom the king greatly feared. At Ormuz in March 1515, Afonso met the king and asked the vizier to be present. He then had him immediately stabbed and killed by his entourage, thus "freeing" the terrified king, so the island in the Persian Gulf yielded to him without resistance and remained a vassal state of the Portuguese Empire. Ormuz itself would not be Persian territory for another century, until an English-Persian alliance finally expelled the Portuguese in 1622. At Ormuz, Afonso met with Miguel Ferreira, returning with rich presents and an ambassador, carrying a letter from the Persian potentate Shah Ismael, inviting Afonso to become a leading lord in Persia. There he remained, engaging in diplomatic efforts, receiving envoys and overseeing the construction of the new fortress, while becoming increasingly ill. His illness was reported as early as September 1515. In November 1515, he embarked on a journey back to Goa. Death At this time, his political enemies at the Portuguese court were planning his downfall. They had lost no opportunity in stirring up the jealousy of King Manuel against him, insinuating that Afonso intended to usurp power in Portuguese India. While on his return voyage from Ormuz in the Persian Gulf, near the harbor of Chaul, he received news of a Portuguese fleet arriving from Europe, bearing dispatches announcing that he was to be replaced by his personal foe, Lopo Soares de Albergaria. Realizing the plot that his enemies had moved against him, profoundly disillusioned, he voiced his bitterness: "Grave must be my sins before the King, for I am in ill favor with the King for love of the men, and with the men for love of the King." Feeling himself near death, he donned the surcoat of the Order of Santiago, of which he was a knight, and drew up his will, appointed the captain and senior officials of Ormuz, and organized a final council with his captains to decide the main matters affecting the Portuguese State of India. He wrote a brief letter to King Manuel, asking him to confer onto his natural son "all of the high honors and rewards" that Afonso had received, and assuring Manuel of his loyalty. On 16 December 1515, Afonso de Albuquerque died within sight of Goa. As his death was known, in the city "great wailing arose", and many took to the streets to witness his body carried on a chair by his main captains, in a procession lit by torches amidst the crowd. Afonso's body was buried in Goa, according to his will, in the Church of Nossa Senhora da Serra (Our Lady of the Hill), which he had been built in 1513 to thank the Madonna for his escape from Kamaran island. That night, the population of Goa, both Hindu and Portuguese, gathered to mourn his death. In Portugal, King Manuel's zigzagging policies continued, still trapped by the constraints of real-time medieval communication between Lisbon and India and unaware that Afonso was dead. Hearing rumours that the Mamluk Sultan of Egypt was preparing a magnificent army at Suez to prevent the conquest of Ormuz, he repented of having replaced Afonso, and in March 1516 urgently wrote to Albergaria to return the command of all operations to Afonso and provide him with resources to face the Egyptian threat. He organized a new Portuguese navy in Asia, with orders that Afonso (if he was still in India), be made commander-in-chief against the Sultan of Cairo's armies. Manuel would afterwards learn that Afonso had died many months earlier, and that his reversed decision had been delivered many months too late. After 51 years, in 1566, his body was moved to Nossa Senhora da Graça church in Lisbon, which was ruined and rebuilt after the 1755 Great Lisbon earthquake. Legacy King Manuel I of Portugal was belatedly convinced of Afonso's loyalty, and endeavoured to atone for his lack of confidence in Afonso by heaping honours upon his son, Brás de Albuquerque (1500–1580), whom he renamed "Afonso" in memory of the father. Afonso de Albuquerque was a prolific writer, having sent numerous letters during his governorship, covering topics from minor issues to major strategies. In 1557 his son published his biography under the title Commentarios do Grande Affonso d'Alboquerque. In 1572, Afonso's actions were described in The Lusiads, the Portuguese main epic poem by Luís Vaz de Camões (Canto X, strophes 40–49). The poet praises his achievements, but has the muses frown upon the harsh rule of his men, of whom Camões was almost a contemporary fellow. In 1934, Afonso was celebrated by Fernando Pessoa in Mensagem, a symbolist epic. In the first part of this work, called "Brasão" (Coat-of-Arms), he relates Portuguese historical protagonists to each of the fields in the Portuguese coat-of-arms, Afonso being one of the wings of the griffin headed by Henry the Navigator, the other wing being King John II. A variety of mango, which was created by Portuguese Jesuits in Goa via grafting techniques, was named in his honour. Numerous homages have been paid to Afonso; he is featured in the Padrão dos Descobrimentos monument; there is a square named after him in Lisbon, which also features a bronze statue, and two Portuguese Navy ships have been named in his honour: the sloop NRP Afonso de Albuquerque (1884) and the warship NRP Afonso de Albuquerque. Titles and honours Captain-Major of the Sea of Arabia 2nd Governor of India 1st Duke of Goa A knight of the Portuguese Order of Saint James of the Sword Fidalgo of the Royal Household Notes References Citations Bibliography In other languages Albuquerque, Afonso de, D. Manuel I, António Baião, "Cartas para el-rei d". Manuel I", Editora Livraria Sá de Costa (1957) Primary sources External links Paul Lunde, The coming of the Portuguese, 2006, Saudi Aramco World Portuguese explorers Explorers of Asia Explorers of India Viceroys of Portuguese India Date of birth unknown 1450s births 1515 deaths Portuguese admirals Portuguese colonial governors and administrators Portuguese generals Portuguese Renaissance writers People from Vila Franca de Xira Colonial Goa Colonial Kerala Maritime history of Portugal Portuguese in Kerala History of Kollam Shipwreck survivors 1510s in Portuguese India 16th-century Portuguese people Portuguese nobility
1694
https://en.wikipedia.org/wiki/Alfonso%20IV
Alfonso IV
Alfonso IV may refer to: Alfonso IV of León (924–931) Afonso IV of Portugal (1291–1357) Alfonso IV of Aragon (1327–1336) Alfonso IV of Ribagorza (1332–1412) Alfonso IV d'Este (1634–1662), Duke of Modena and Regg
1828
https://en.wikipedia.org/wiki/Amitabh%20Bachchan
Amitabh Bachchan
Amitabh Bachchan (; born as Amitabh Shrivastav; 11 October 1942) is an Indian actor, film producer, television host, occasional playback singer and former politician, who works in Hindi cinema. In film career spanning over five decades, he has starred in more than 200 films. Bachchan is widely regarded as one of the most successful and influential actors in the history of Indian cinema. Referred to as the Shahenshah of Bollywood, Sadi Ke Mahanayak (Hindi for, "Greatest actor of the century"), Star of the Millennium, or Big B. His dominance in the Indian movie scene during the 1970s1980s made the French director François Truffaut call it a "one-man industry". Bachchan was born in 1942 in Allahabad to the Hindi poet Harivansh Rai Bachchan and his wife, the social activist Teji Bachchan. He was educated at Sherwood College, Nainital, and Kirori Mal College, University of Delhi. His film career started in 1969 as a voice narrator in Mrinal Sen's film Bhuvan Shome. He first gained popularity in the early 1970s for films such as Zanjeer, Deewaar and Sholay, and achieved greater stardom in later years, dubbed India's "angry young man" for several of his on-screen roles in Hindi films. He consistently starred in top grossing Indian films with critical acclaim since mid 1970s to 80s, such as Amar Akbar Anthony, Don, Trishul, Muqaddar Ka Sikander, Kaala Patthar, Kaalia, Naseeb, Namak Halaal, Coolie, Sharaabi and Mard, as well as some of his most acclaimed performances include Namak Haraam, Abhimaan, Mili, Silsala, Dostana, Shakti, Shahenshah and Agneepath. After taking break from acting in 1990s, his resurgence marked in 2000 with Mohabbatein. Since then he starred in several successful and acclaimed films such as Aankhen, Baghban, Black, Sarkar, Paa, Bhoothnath and Piku; for which he won his forth National Film Award for Best Actor, making him the only actor to do so. Bachchan also made an appearance in a Hollywood film, The Great Gatsby (2013), in which he played a non-Indian Jewish character. He has won numerous accolades in his career, including four National Film Awards as Best Actor and many awards at international film festivals and award ceremonies. He has won sixteen Filmfare Awards and is the most nominated performer in any major acting category at Filmfare with 34 nominations in Best Actor and 42 nominations overall. The Government of India honoured him with the Padma Shri in 1984, the Padma Bhushan in 2001, the Padma Vibhushan in 2015 and India's highest award in the field of cinema, Dadasaheb Phalke Award in 2018 for his contributions to the arts. The Government of France honoured him with its highest civilian honour, Knight of the Legion of honour, in 2007 for his exceptional career in the world of cinema and beyond. In addition to acting, Bachchan has worked as a playback singer, film producer and television presenter. He has hosted several seasons of the game show Kaun Banega Crorepati, India's version of the game show franchise, Who Wants to Be a Millionaire?. He also entered politics for a time in the 1980s. Bachchan has also been involved in several humanitarian works and he is a leading brand endorser in India. Beyond the Indian subcontinent, he acquired a large overseas following of the South Asian diaspora, as well as others, in markets including Africa (South Africa, Eastern Africa and Mauritius), the Middle East (especially UAE and Egypt), the United Kingdom, Russia, Central Asia, the Caribbean (Guyana, Suriname, and Trinidad and Tobago), Oceania (Fiji, Australia, and New Zealand), Canada and the United States. Early life and family Bachchan was born on 11 October 1942 in Allahabad to the Hindi poet Harivansh Rai Bachchan, and social activist Teji Bachchan. Harivansh Rai Bachchan was an Awadhi Hindu Kayastha, who was fluent in Awadhi, Hindi and Urdu. Harivansh's ancestors came from a village called Babupatti, in the Raniganj tehsil, in the Pratapgarh district, in the present-day state of Uttar Pradesh, in India. Teji Bachchan was a Punjabi Sikh Khatri from Lyallpur, Punjab, British India (present-day Faisalabad, Punjab, Pakistan). Bachchan has a younger brother, Ajitabh, he is 5 years younger than him. Bachchan's parents were initially going to name him Inquilaab (Hindustani for "Revolution"), inspired by the phrase Inquilab Zindabad (which translates into English as "Long live the revolution") popularly used during the Indian independence struggle; the name Amitabh was suggested to his father by poet Sumitranandan Pant. Although his surname was Shrivastava, Amitabh's father, who opposed the caste system, had adopted the pen name Bachchan ("child-like" in colloquial Hindi), under which he published all of his works. When his father was looking to get him admitted to a school, he and Bachchan's mother decided the family's name should be Bachchan instead of Shrivastava. It is with this last name that Amitabh debuted in films and used for all other practical purposes, Bachchan has become the surname for all of his immediate family. Bachchan's father died in 2003, and his mother in 2007. Bachchan's secondary education was at Boys' High School & College in Allahabad and Sherwood College in Nainital. He attended Kirori Mal College at the University of Delhi in Delhi. He graduated with Bachelor of Science degree from Kirori Mal College in 1962. When Bachchan finished his studies his father approached Prithviraj Kapoor, the founder of Prithvi Theatre and patriarch of the Kapoor acting family, to see if there was an opening for him, but Kapoor offered no encouragement. Bachchan was a friend of Rajiv Gandhi and Sanjay Gandhi, before he became an actor. He used to spend time with them when he was a resident in New Delhi. Bachchan's family were very close to the Nehru-Gandhi family of politicians. When Sonia Gandhi first came to India from Italy before her marriage, Bachchan had received her at the Palam International Airport on 13 January 1968. She spent 48 days at Bachchan's house with his parents before her marriage to Rajiv. Bachchan applied for a role as a newsreader for All India Radio, Delhi but "failed the audition". He became a business executive for Bird & Company in Kolkata (Calcutta), and worked in the theatre before starting his film career. It is thought that his mother might have had some influence in Amitabh Bachchan's choice of career because she always insisted that he should "take centre stage". Acting career Early career (1969–1972) Bachchan made his film debut in 1969, as a voice narrator in Mrinal Sen's National Award-winning film Bhuvan Shome. His first acting role was as one of the seven protagonists in the film Saat Hindustani, directed by Khwaja Ahmad Abbas and featuring Utpal Dutt, Anwar Ali (brother of comedian Mehmood), Madhu and Jalal Agha. Anand (1971) followed, in which Bachchan starred alongside Rajesh Khanna. His role as a doctor with a cynical view of life garnered Bachchan his first Filmfare Award for Best Supporting Actor. He then played his first antagonist role as an infatuated lover-turned-murderer in Parwana (1971). Following Parwana were several films including Reshma Aur Shera (1971). During this time, he made a guest appearance in the film Guddi which starred his future wife Jaya Bhaduri. He narrated part of the film Bawarchi. In 1972, he made an appearance in the road action comedy Bombay to Goa directed by S. Ramanathan which was moderately successful. Many of Bachchan's films during this early period did not do well. His only film with Mala Sinha, Sanjog (1972) was also a box office failure. Rise to stardom (1973–1974) Bachchan was struggling, seen as a "failed newcomer" who, by the age of 30, had twelve flops and only two hits (as a lead in Bombay to Goa and supporting role in Anand). He was offered with a dual role movie by the director O.P Goyle, and writer O.P Ralhan for the film Bandhe Hath in 1973. This was Bachchan's first movie where he had played double role. Bachchan was soon discovered by screenwriter duo Salim–Javed, consisting of Salim Khan and Javed Akhtar. Salim Khan wrote the story, screenplay and script of Zanjeer (1973), and conceived the "angry young man" persona of the lead role. Javed Akhtar came on board as co-writer, and Prakash Mehra, who saw the script as potentially groundbreaking, as the film's director. However, they were struggling to find an actor for the lead "angry young man" role; it was turned down by a number of actors, owing to it going against the "romantic hero" image dominant in the industry at the time. Salim-Javed soon discovered Bachchan and "saw his talent, which most makers didn't. He was exceptional, a genius actor who was in films that weren't good." According to Salim Khan, they "strongly felt that Amitabh was the ideal casting for Zanjeer". Salim Khan introduced Bachchan to Prakash Mehra, and Salim-Javed insisted that Bachchan be cast for the role. Zanjeer was a crime film with violent action, in sharp contrast to the romantically themed films that had generally preceded it, and it established Amitabh in a new persona—the "angry young man" of Bollywood cinema. He earned his first Filmfare Award nomination for Best Actor, with Filmfare later considering this one of the most iconic performances of Bollywood history. The film was a huge success and one of the highest-grossing films of that year, breaking Bachchan's dry spell at the box office and making him a star. It was the first of many collaborations between Salim-Javed and Amitabh Bachchan; Salim-Javed wrote many of their subsequent scripts with Bachchan in mind for the lead role, and insisted on him being cast for their later films, including blockbusters such as Deewaar (1975) and Sholay (1975). Salim Khan also introduced Bachchan to director Manmohan Desai with whom he formed a long and successful association, alongside Prakash Mehra and Yash Chopra. Eventually, Bachchan became one of the most successful leading men of the film industry. Bachchan's portrayal of the wronged hero fighting a crooked system and circumstances of deprivation in films like Zanjeer, Deewaar, Trishul, Kaala Patthar and Shakti resonated with the masses of the time, especially the youth who harboured a simmering discontent owing to social ills such as poverty, hunger, unemployment, corruption, social inequality and the brutal excesses of The Emergency. This led to Bachchan being dubbed as the "angry young man", a journalistic catchphrase which became a metaphor for the dormant rage, frustration, restlessness, sense of rebellion and anti-establishment disposition of an entire generation, prevalent in 1970s India. The year 1973 was also when he married Jaya, and around this time they appeared in several films together: not only Zanjeer but also subsequent films such as Abhimaan, which was released only a month after their marriage and was also successful at the box office. Later, Bachchan played the role of Vikram, once again along with Rajesh Khanna, in the film Namak Haraam, a social drama directed by Hrishikesh Mukherjee and scripted by Biresh Chatterjee addressing themes of friendship. His supporting role won him his second Filmfare Award for Best Supporting Actor. In 1974, Bachchan made several guest appearances in films such as Kunwara Baap and Dost, before playing a supporting role in Roti Kapda Aur Makaan. The film, directed and written by Manoj Kumar, addressed themes of honesty in the face of oppression and financial and emotional hardship and was the top-earning film of 1974. Bachchan then played the leading role in the film Majboor. The film was a success at the box office. Superstardom (1975–1988) In 1975, he starred in a variety of film genres, from the comedy Chupke Chupke and the crime drama Faraar to the romantic drama Mili. This was also the year in which Bachchan starred in two films regarded as important in Hindi cinema history, both written by Salim-Javed, who again insisted on casting Bachchan. The first was Deewaar, directed by Yash Chopra, where he worked with Shashi Kapoor, Nirupa Roy, Parveen Babi, and Neetu Singh, and earned another Filmfare nomination for Best Actor. The film became a major hit at the box office in 1975, ranking in at number four. Indiatimes Movies ranks Deewaar amongst the Top 25 Must See Bollywood Films. The other, released on 15 August 1975, was Sholay, which became the highest-grossing film ever in India at the time, in which Bachchan played the role of Jaidev. Deewaar and Sholay are often credited with exalting Bachchan to the heights of superstardom, two years after he became a star with Zanjeer, and consolidating his domination of the industry throughout the 1970s and 1980s. In 1999, BBC India declared Sholay the "Film of the Millennium" and, like Deewaar, it has been cited by Indiatimes Movies as amongst the Top 25 Must See Bollywood Films. In that same year, the judges of the 50th annual Filmfare Awards awarded it with the special distinction award called the Filmfare Best Film of 50 Years. In 1976, he was cast by Yash Chopra in the romantic family drama Kabhie Kabhie. Bachchan starred as a young poet, Amit Malhotra, who falls deeply in love with a beautiful young girl named Pooja (Rakhee Gulzar) who ends up marrying someone else (Shashi Kapoor). The film was notable for portraying Bachchan as a romantic hero, a far cry from his "angry young man" roles like Zanjeer and Deewaar. The film evoked a favourable response from critics and audiences alike. Bachchan was again nominated for the Filmfare Best Actor Award for his role in the film. That same year he played a double role in the hit Adalat as father and son. In 1977, he won his first Filmfare Best Actor Award for his performance in Amar Akbar Anthony, in which he played the third lead opposite Vinod Khanna and Rishi Kapoor as Anthony Gonsalves. The film was the highest-grossing film of that year. His other successes that year include Parvarish and Khoon Pasina. He once again resumed double roles in films such as Kasme Vaade (1978) as Amit and Shankar and Don (1978) playing the characters of Don, a leader of an underworld gang and his look-alike Vijay. His performance won him his second Filmfare Best Actor Award. He also gave towering performances in Yash Chopra's Trishul and Prakash Mehra's Muqaddar Ka Sikandar both of which earned him further Filmfare Best Actor nominations. 1978 is arguably considered his most successful year at the box office since all of his six releases the same year, namely Muqaddar Ka Sikandar, Trishul, Don, Kasme Vaade, Ganga Ki Saugandh and Besharam were massive successes, the former three being the consecutive highest-grossing films of the year, a rare feat in Indian cinema. In 1979, Bachchan starred in Suhaag which was the highest earning film of that year. In the same year he also enjoyed critical acclaim and commercial success with films like Mr. Natwarlal, Kaala Patthar, The Great Gambler and Manzil. Amitabh was required to use his singing voice for the first time in a song from the film Mr. Natwarlal in which he starred with Rekha. Bachchan's performance in the film saw him nominated for both the Filmfare Best Actor Award and the Filmfare Award for Best Male Playback Singer. He also received Best Actor nomination for Kaala Patthar and then went on to be nominated again in 1980 for the Raj Khosla directed film Dostana, in which he starred opposite Shatrughan Sinha and Zeenat Aman. Dostana proved to be the top-grossing film of 1980. In 1981, he starred in Yash Chopra's melodrama film Silsila, where he starred alongside his wife Jaya and also Rekha. Other successful films of this period include Shaan (1980), Ram Balram (1980), Naseeb (1981), Lawaaris (1981), Kaalia (1981), Yaarana (1981), Barsaat Ki Ek Raat (1981) and Shakti (1982), also starring Dilip Kumar. In 1982, he played double roles in the musical Satte Pe Satta and action drama Desh Premee which succeeded at the box office along with mega hits like action comedy Namak Halaal, action drama Khud-Daar and the critically acclaimed drama Bemisal. In 1983, he played a triple role in Mahaan which was not as successful as his previous films. Other releases during that year included Nastik and Pukar which were hits and Andha Kanoon (in which he had an extended guest appearance) was an average grosser. During a stint in politics from 1984 to 1987, his completed films Mard (1985) and Aakhree Raasta (1986) were released and were major hits. Bachchan had played a role in a special appearance for the movie Kaun Jeeta Kaun Haara in the year 1987 and he sang a playback song with Kishore Kumar in this movie. Coolie incident On 26 July 1982, while filming a fight scene with co-actor Puneet Issar for Coolie, Bachchan had near-fatal intestinal injury. Bachchan was performing his own stunts in the film and one scene required him to fall onto a table and then on the ground. However, as he jumped towards the table, the corner of the table struck his abdomen, resulting in a splenic rupture from which he lost a significant amount of blood. He required an emergency splenectomy and remained critically ill in hospital for many months, at times close to death. There were long queues of well-wishing fans outside the hospital where he was recuperating; the public response included prayers in temples and offers to sacrifice limbs to save him. Nevertheless, he resumed filming later that year after a long period of recuperation. The director, Manmohan Desai, altered the ending of Coolie: Bachchan's character was originally intended to have been killed off; but, after the change of script, the character lived in the end. Desai felt it would have been inappropriate for the man who had just fended off death in real life to be killed on screen. The footage of the fight scene is frozen at the critical moment, and a caption appears onscreen marking it as the instant of the actor's injury. The film was released in 1983, and partly due to the huge publicity of Bachchan's accident, the film was a box office success and the top-grossing film of that year. Health issues Later, he was diagnosed with Myasthenia gravis. His illness made him feel weak both mentally and physically and he decided to quit films and venture into politics. At this time he became pessimistic, expressing concern with how a new film would be received, and stating before every release, "Yeh film to flop hogi!" ("This film will flop"). Career fluctuations and sabbatical (1988–1992) After a three-year stint in politics from 1984 to 1987, Bachchan returned to films in 1988, playing the title role in Shahenshah, which was a box office success. After the success of his comeback film however, his star power began to wane as all of his subsequent films like Jaadugar, Toofan and Main Azaad Hoon (all released in 1989) failed at the box office. He gained success during this period with the crime drama Aaj Ka Arjun (1990) and action crime drama Hum (1991), for which he won his third Filmfare Best Actor Award, but this momentum was short-lived and his string of box office failures continued. Notably, despite the lack of hits, it was during this era that Bachchan won his first National Film Award for Best Actor for his performance as a Mafia don in the 1990 cult film Agneepath. These years would see his last on-screen appearances for some time. After the release of the critically acclaimed epic Khuda Gawah in 1992, Bachchan went into semi-retirement for five years. With the exception of the delayed release of Insaniyat (1994), which was also a box office failure, Bachchan did not appear in any new releases for five years. Business ventures and acting comeback (1996–1999) Bachchan turned producer during his temporary retirement period, setting up Amitabh Bachchan Corporation, Ltd. (ABCL) in 1996. ABCL's strategy was to introduce products and services covering an entire cross-section of India's entertainment industry. ABCL's operations were mainstream commercial film production and distribution, audio cassettes and video discs, production and marketing of television software, and celebrity and event management. Soon after the company was launched in 1996, the first film it produced was Tere Mere Sapne, which was a moderate success and launched the careers of actors like Arshad Warsi and southern film star Simran. In 1997, Bachchan attempted to make his acting comeback with the film Mrityudata, produced by ABCL. Though Mrityudaata attempted to reprise Bachchan's earlier success as an action hero, the film was a failure both financially and critically. ABCL was the main sponsor of the 1996 Miss World beauty pageant, Bangalore, but lost millions. The fiasco and the consequent legal battles surrounding ABCL and various entities after the event, coupled with the fact that ABCL was reported to have overpaid most of its top-level managers, eventually led to its financial and operational collapse in 1997. The company went into administration and was later declared a failed company by the Indian Industries board. The Bombay high court, in April 1999, restrained Bachchan from selling off his Bombay bungalow 'Prateeksha' and two flats until the pending loan recovery cases of Canara Bank were disposed of. Bachchan had, however, pleaded that he had mortgaged his bungalow to raise funds for his company. Bachchan attempted to revive his acting career, and eventually had commercial success with Bade Miyan Chote Miyan (1998) and Major Saab (1998), and received positive reviews for Sooryavansham (1999), but other films such as Lal Baadshah (1999) and Hindustan Ki Kasam (1999) were box office failures. Return to prominence (2000–present) In 2000, Bachchan appeared in Yash Chopra's box-office hit, Mohabbatein, directed by Aditya Chopra. He played a stern, elder figure who rivalled the character of Shahrukh Khan. His role won him his third Filmfare Award for Best Supporting Actor. Other hits followed, with Bachchan appearing as an older family patriarch in Ek Rishtaa: The Bond of Love (2001), Kabhi Khushi Kabhie Gham... (2001) and Baghban (2003). As an actor, he continued to perform in a range of characters, receiving critical praise for his performances in Aks (2001), Aankhen (2002), Kaante (2002), Khakee (2004) and Dev (2004). His performance in Aks won him his first Filmfare Critics Award for Best Actor. One project that did particularly well for Bachchan was Sanjay Leela Bhansali's Black (2005). The film starred Bachchan as an ageing teacher of a deaf-blind girl and followed their relationship. His performance was unanimously praised by critics and audiences and won him his second National Film Award for Best Actor, his fourth Filmfare Best Actor Award and his second Filmfare Critics Award for Best Actor. Taking advantage of this resurgence, Amitabh began endorsing a variety of products and services, appearing in many television and billboard advertisements. In 2005 and 2006, he starred with his son Abhishek in the films Bunty Aur Babli (2005), the Godfather tribute Sarkar (2005), and Kabhi Alvida Naa Kehna (2006). All of them were successful at the box office. His later releases in 2006 and early 2007 were Baabul (2006), Ekalavya and Nishabd (2007), which failed to do well at the box office but his performances in each of them were praised by critics. In May 2007, two of his films: the romantic comedy Cheeni Kum and the multi-starrer action drama Shootout at Lokhandwala were released. Shootout at Lokhandwala did well at the box office and was declared a hit in India, while Cheeni Kum picked up after a slow start and was a success. A remake of his biggest hit, Sholay (1975), entitled Ram Gopal Varma Ki Aag, released in August of that same year and proved to be a major commercial failure in addition to its poor critical reception. The year also marked Bachchan's first appearance in an English-language film, Rituparno Ghosh's The Last Lear, co-starring Arjun Rampal and Preity Zinta. The film premiered at the 2007 Toronto International Film Festival on 9 September 2007. He received positive reviews from critics who hailed his performance as his best ever since Black. Bachchan was slated to play a supporting role in his first international film, Shantaram, directed by Mira Nair and starring Hollywood actor Johnny Depp in the lead. The film was due to begin filming in February 2008 but due to the writer's strike, was pushed to September 2008. The film is currently "shelved" indefinitely. Vivek Sharma's Bhoothnath, in which he plays the title role as a ghost, was released on 9 May 2008. Sarkar Raj, the sequel of the 2005 film Sarkar, released in June 2008 and received a positive response at the box office. Paa, which released at the end of 2009 was a highly anticipated project as it saw him playing his own son Abhishek's Progeria-affected 13-year-old son, and it opened to favourable reviews, particularly towards Bachchan's performance and was one of the top-grossing films of 2009. It won him his third National Film Award for Best Actor and fifth Filmfare Best Actor Award. In 2010, he debuted in Malayalam film through Kandahar, directed by Major Ravi and co-starring Mohanlal. The film was based on the hijacking incident of the Indian Airlines Flight 814. Bachchan declined any remuneration for this film. In 2011 he played an aged retired former gangster in Bbuddah... Hoga Terra Baap who protects his son Sonu Sood who is an honest daring police officer from a notorious gangster Prakash Raj who unknowingly hired the latter to perform a contract killing not knowing that the police officer is the gangster's son. Directed By Puri Jagannadh the film won positive reviews and was a commercial success. In 2013 he made his Hollywood debut in The Great Gatsby making a special appearance opposite Leonardo DiCaprio and Tobey Maguire. In 2014, he played the role of the friendly ghost in the sequel Bhoothnath Returns. The next year, he played the role of a grumpy father experiencing chronic constipation in the critically acclaimed Piku which was also one of the biggest hits of 2015. A review in Daily News and Analysis (DNA) summarised Bachchan's performance as "The heart and soul of Piku clearly belong to Amitabh Bachchan who is in his elements. His performance in Piku, without doubt, finds a place among the top 10 in his illustrious career." Rachel Saltz wrote for The New York Times, "Piku", an offbeat Hindi comedy, would have you contemplate the intestines and mortality of one Bhashkor Banerji and the actor who plays him, Amitabh Bachchan. Bhashkor's life and conversation may revolve around his constipation and fussy hypochondria, but there's no mistaking the scene-stealing energy that Mr. Bachchan, India's erstwhile Angry Young Man, musters for his new role of Cranky Old Man." Well known Indian critic Rajeev Masand wrote on his website, "Bachchan is pretty terrific as Bhashkor, who reminds you of that oddball uncle that you nevertheless have a soft spot for. He bickers with the maids, harrows his hapless helper, and expects that Piku stay unmarried so she can attend to him. At one point, to ward off a possible suitor, he casually mentions that his daughter isn't a virgin; that she's financially independent and sexually independent too. Bachchan embraces the character's many idiosyncrasies, never once slipping into caricature while all along delivering big laughs thanks to his spot-on comic timing." The Guardian summed up, "Bachchan seizes upon his cranky character part, making Bashkor as garrulously funny in his theories on caste and marriage as his system is backed-up." The performance won Bachchan his fourth National Film Award for Best Actor and his third Filmfare Critics Award for Best Actor. In 2016, he appeared in the women-centric courtroom drama film Pink which was highly praised by critics and with an increasingly good word of mouth, was a resounding success at the domestic and overseas box office. Bachchan's performance in the film received acclaim. According to Raja Sen of Rediff.com, "Amitabh Bachchan, a retired lawyer with bipolar disorder, takes up cudgels on behalf of the girls, delivering courtroom blows with pugilistic grace. Like we know from Prakash Mehra movies, into each life some Bachchan must fall. The girls hang on to him with incredulous desperation, and he bats for them with all he has. At one point Meenal hangs by Bachchan's elbow, words entirely unnecessary. Bachchan towers through Pink – the way he bellows "et cetera" is alone worth having the heavy-hitter at play—but there are softer moments like one where he appears to have dozed off in court, or where he lays his head by his convalescent wife's bedside and needs his hair ruffled and his conviction validated." Writing for Hindustan Times, noted film critic and author Anupama Chopra said of Bachchan's performance, "A special salute to Amitabh Bachchan, who imbues his character with a tragic majesty. Bachchan towers in every sense, but without a hint of showboating. Meena Iyer of The Times of India wrote, "The performances are pitch-perfect with Bachchan leading the way. Writing for NDTV, Troy Ribeiro of Indo-Asian News Service (IANS) stated, 'Amitabh Bachchan as Deepak Sehgall, the aged defence lawyer, shines as always, in a restrained, but powerful performance. His histrionics come primarily in the form of his well-modulated baritone, conveying his emotions and of course, from the well-written lines.' Mike McCahill of The Guardian remarked, "Among an electric ensemble, Tapsee Pannu, Kirti Kulhari and Andrea Tariang give unwavering voice to the girls' struggles; Amitabh Bachchan brings his moral authority to bear as their sole legal ally. In 2017, he appeared in the third instalment of the Sarkar film series: Ram Gopal Varma's Sarkar 3. That year, he started filming for the swashbuckling action adventure film Thugs Of Hindostan with Aamir Khan, Katrina Kaif and Fatima Sana Shaikh which released in November 2018. He co-starred with Rishi Kapoor in 102 Not Out, a comedy drama film directed by Umesh Shukla based on a Gujarati play of the same name written by Saumya Joshi. This film released in May 2018 and reunited him with Kapoor onscreen after a gap of twenty-seven years. In 2019, he played the role of Badal Gupta in Sujoy Ghosh's Badla. Later that year, he made his Telugu debut in Surender Reddy's Sye Raa Narasimha Reddy as Gosayi Venkanna. He did Gulabo Sitabo in 2020 for which he received Filmfare Critics Award For Best Actor. In 2021, he appeared in Rumy Jaffery's mystery thriller Chehre along with Emraan Hashmi. In 2022, he did 5 films: Jhund, Runway 34, Brahmāstra: Part One – Shiva, Goodbye and Uunchai. Now he is all set to appear in Nag Ashwin's Project K. Other work Television appearances In 2000, Bachchan hosted the first season of Kaun Banega Crorepati (KBC), the Indian adaptation of the British television game show, Who Wants to Be a Millionaire?. The show was well received. A second season followed in 2005 but its run was cut short by Star Plus when Bachchan fell ill in 2006. In 2009, Bachchan hosted the third season of the reality show Bigg Boss. In 2010, Bachchan hosted the fourth season of KBC. The fifth season started on 15 August 2011 and ended on 17 November 2011. The show became a massive hit with audiences and broke many TRP Records. CNN IBN awarded Indian of the Year- Entertainment to Team KBC and Bachchan. The Show also grabbed all the major Awards for its category. The sixth season was also hosted by Bachchan, commencing on 7 September 2012, broadcast on Sony TV and received the highest number of viewers thus far. In 2014, he debuted in the fictional Sony Entertainment Television TV series titled Yudh playing the lead role of a businessman battling both his personal and professional life. Voice-acting Bachchan is known for his deep, baritone voice. He has been a narrator, a playback singer, and presenter for numerous programmes. Some prominent films featuring his narration are Satyajit Ray's 1977 film Shatranj Ke Khiladi. and Ashutosh Gowarikar's 2001 film Lagaan. He also has done voice-over work for the following movies: Bhuvan Shome (1969) Bawarchi (1972) Balika Badhu (1975) Tere Mere Sapne (1996) Hello Brother (1999) Lagaan (2001) Fun2shh... Dudes in the 10th Century (2003) Parineeta (2005) March of the Penguins (2005), Indian version Jodhaa Akbar (2008) Swami (2007) Zor Lagaa Ke...Haiya! (2009) Ra.One (2011) Kahaani (2012) Krrish 3 (2013) Mahabharat (2013) Kochadaiiyaan (Hindi Version) (2014) CBI documentary (2014) – sanctioned by Central Bureau of Investigation The Ghazi Attack (2017) Firangi (2017) Business investments Around 1994, Bachchan started Amitabh Bachchan Corporation Ltd (ABCL), an event management, production and distribution company. But the company led into debt with fiasco and gone into bankruptcy, subsequently Bachchan became nearly bankrupt. Reasons of this debacle was flop films such as Mrityudata, Major Saab (produced by this organisation), Miss World 1996 which was organised-managed by ABCL. Due to this he began work for TV, asked for work to Yash Chopra. Once he told that, 'it was darkest time for him'. He has invested in many upcoming business ventures. In 2013, he bought a 10% stake in Just Dial from which he made a gain of 4600 per cent. He holds a 3.4% equity in Stampede Capital, a financial technology firm specialising in cloud computing for financial markets. The Bachchan family also bought shares worth $252,000 in Meridian Tech, a consulting company in U.S. Recently they made their first overseas investment in Ziddu.com, a cloud based content distribution platform. Bachchan was named in the Panama Papers and Paradise Papers, leaked confidential documents relating to offshore investment. Political career In 1984, Bachchan took a break from acting and briefly entered politics in support of a long-time family friend, Rajiv Gandhi. He contested Allahabad's seat for the 8th Lok Sabha against H. N. Bahuguna, former Chief Minister of Uttar Pradesh. With 68.2% of the votes in his favour, he won by one of the highest victory margins ever in Indian elections. In 1987, Indian Express said his brother Ajitabh Bachchan owned an apartment in Switzerland, giving rise to speculations about his involvement in the "Bofors scandal", revealed in the year before. Bachchan resigned from his seat in July 1987. Ajitabh Bachchan sued Swedish newspaper Dagens Nyheter for linking him to Bofors payments in 1990 and won damages in the United Kingdom. Sten Lindstrom, the Swedish police chief who had investigated the case, said in 2012 that "Indian investigators planted the Bachchan angle on" Dagens Nyheter. Bachchan's old friend, Amar Singh, helped him during the financial crisis caused by the failure of his company, ABCL. Thereafter Bachchan started supporting the Samajwadi Party, the political party to which Amar Singh belonged. Jaya Bachchan joined the Samajwadi Party and represented the party as an MP in the Rajya Sabha. Bachchan appeared in advertisements and political campaigns for the party. His claim that he too was a farmer in the advertisements were questioned in courts. Bachchan has claimed to have been banned by film press during the emergency years for his family's friendship with Indira Gandhi. Bachchan has been accused of using the slogan "blood for blood" in the context of the 1984 anti-Sikh riots. Bachchan has denied the allegation. In October 2014, Bachchan was summoned by a court in Los Angeles for "allegedly instigating violence against the Sikh community". Bachchan in an interview with journalist Arnab Goswami offered to fight the case in court and asked the accusers to file the same as also present proof. He was also one of the trustees of Rajiv Gandhi Foundation. Humanitarian and social causes Bachchan has been involved with many social causes. For example, he donated to clear the debts of nearly 40 beleaguered farmers in Andhra Pradesh and to clear the debts of some 100 Vidarbha farmers. In 2010, he donated to Resul Pookutty's foundation for a medical centre at Kochi, and he has given () to the family of Delhi policeman Subhash Chand Tomar who died after succumbing to injuries during a protest against gang-rape after the 2012 Delhi gang rape case. He founded the Harivansh Rai Bachchan Memorial Trust, named after his father, in 2013. This trust, in association with Urja Foundation, will be powering 3,000 homes in India with electricity through solar energy. In June 2019 he cleared debts of 2100 farmers from Bihar. Bachchan was made a UNICEF goodwill ambassador for the polio Eradication Campaign in India in 2002. In 2013, he and his family donated () to a charitable trust, Plan India, that works for the betterment of young girls in India. He also donated () to the Maharashtra Police Welfare Fund in 2013. Bachchan was the face of the 'Save Our Tigers' campaign that promoted the importance of tiger conservation in India. He supported the campaign by PETA in India to free Sunder, a 14-year-old elephant who was chained and tortured in a temple in Kolhapur, Maharashtra. In 2014, it was announced that he had recorded his voice and lent his image to the Hindi and English language versions of the TeachAids software, an international HIV/AIDS prevention education tool developed at Stanford University. He has been a vocal "brand ambassador" of the Swachh Bharat Mission (SBM) and featured in a few advertisements to promote the campaign. In 2020, Bachchan was helping the Government of India promote its public health message concerning COVID-19 before he and some members of his family themselves became infected. He was hospitalised with reported mild symptoms of the disease on 11 July. He was discharged from hospital on 2 August. During the pandemic he lent his support by donating Oxygen concentrators and 25 cr rupees in various forms. Personal life Bachchan has been married to veteran actress and politician Jaya Bhaduri since June 3, 1973, when he was 30 years old, and together they have two children; Abhishek, an actor, and Shweta, an author, journalist and former model. Abhishek married actress Aishwarya Rai, and they have a daughter named Aaradhya. Shweta is married to businessman Nikhil Nanda who is a part of the Kapoor family of actors. They have a daughter, Navya Naveli, and a son, Agastya. Amitabh's family lives in Mumbai in Maharashtra. His younger brother Ajitabh Bachchan is a businessman. He did business and lived in London for brief period of time. Presently he is living in India. He and his family choose to stay away from limelight. His wife Ramola is a fashion designer and was active in business. Ajitabh has one son, Bhim, and three daughters Naina, Namrata and Nilima. Naina Bachchan is married to actor Kunal Kapoor. Bachchan was famously rumoured to have had an extramarital affair with actress Rekha in the mid-1970s to the early 1980s after they first acted together in Do Anjaane, and later in many successful films like Khoon Pasina, Ganga Ki Saugandh, Muqaddar Ka Sikandar, Mr. Natwarlal, Suhaag, Ram Balram and ending in Silsila, though they have both denied it. Filmography Legacy Bachchan is widely regarded as one of the greatest and most influential actors in the history of Indian cinema. He earned respect among critics for his memorable performances and charismatic screen presence, is also considered one of the most respected public figure of India. Referred to as the "Shahenshah of Bollywood", "Star of the Millennium" or "Big B", He inspired many great and successful Indian cinema actors for many generations including Rajinikanth, Chiranjeevi, Kamal Hassan, Shah Rukh Khan, Akshay Kumar, Manoj Bajpayee, Ajay Devgn, Mohanlal, Ranveer Singh, Allu Arjun and Yash. Referred to as the "Shahenshah of Bollywood", "Star of the Millennium" or "Big B", French director François Truffaut called him a "one-man industry." In 1999, Bachchan was voted the "greatest star of stage or screen" in a BBC Your Millennium online poll. The organisation noted that "Many people in the western world will not have heard of [him] ... [but it] is a reflection of the huge popularity of Indian films." In October 2003, TIME magazine dubbed Bachchan as "the Undisputed Godfather of Bollywood". In April 2005, The Walter Reade Theater of Lincoln Center in New York honored Bachchan with a special tribute, retrospective—titled "Amitabh Bachchan: The Biggest Film Star in the World". In the early 80s, Bachchan authorised the use of his likeness for the comic book character Supremo in a series titled The Adventures of Amitabh Bachchan. In May 2014, La Trobe University in Australia named a Scholarship after Bachchan. In June 2000, he became the first living Asian to have been modelled in wax at London's Madame Tussauds Wax Museum. Another statue was installed in New York in 2009, Hong Kong in 2011, Bangkok in 2011, Washington, DC in 2012 and Delhi in 2017. In March 2010, Bachchan has been named the list of CNN's "top 25 Asian actors of all time". He was named "Hottest Vegetarian male" by PETA India in 2012. He also won the title of "Asia's Sexiest Vegetarian male" in a contest poll run by PETA Asia in 2008. In Allahabad, the Amitabh Bachchan Sports Complex and Amitabh Bachchan Road are named after him. A government senior secondary school in Saifai, Etawah is called Amitabh Bachchan Government Inter College. There is a waterfall in Sikkim known as Amitabh Bachchan Falls. In 2022, on the occasion of Bachchan's 80th birthday, not-for-profit organisation Film Heritage Foundation announced a film festival as a part of his 11 films collection had screened in 17 cities across the country shown in limited movies theatres. Biographies Several books have been written about Bachchan. The following is the listing of books focused on his life career: Amitabh Bachchan: the Legend was published in 1999, To be or not to be: Amitabh Bachchan in 2004, AB: The Legend (A Photographer's Tribute) in 2006, Amitabh Bachchan: Ek Jeevit Kimvadanti in 2006, Amitabh: The Making of a Superstar in 2006, Looking for the Big B: Bollywood, Bachchan and Me in 2007 and Bachchanalia in 2009. Awards and honours Apart from industry awards won for his performances throughout the years, Bachchan has received several honours for his achievements in the Indian film industry. In 1991, he became the first artist to receive the Filmfare Lifetime Achievement Award, which was established in the name of Raj Kapoor. Bachchan was crowned as Superstar of the Millennium in 2000 at the Filmfare Awards. In 2001, he was honoured with the Actor of the Century award at the Alexandria International Film Festival in Egypt in recognition of his contribution to the world of cinema. Many other honours for his achievements were conferred upon him at several International Film Festivals, including the Lifetime Achievement Award at the 2010 Asian Film Awards. In 2003, he was conferred with the Honorary Citizenship of the French town of Deauville. The Government of India awarded him with the Padma Shri in 1984, the Padma Bhushan in 2001, the Padma Vibhushan in 2015 and Dadasaheb Phalke Award in 2019. The then-President of Afghanistan awarded him the Order of Afghanistan in 1991 following the shooting of Khuda Gawah there. The Government of Madhya Pradesh honoured him with Rashtriya Kishore Kumar Samman for 2002–2003. France's highest civilian honour, the Knight of the Legion of honour, was conferred upon him by the French Government in 2007 for his "exceptional career in the world of cinema and beyond". On 27 July 2012, at the age of 69, Bachchan carried the Olympic torch during the last leg of its relay in London's Southwark. Bibliography Soul Curry for You and Me – An Empowering Philosophy That Can Enrich Your Life. (2002) See also List of Bollywood actors Lists of Indian actors References Further reading External links 1942 births Officers of the Legion of Honour Male actors in Hindi cinema Male actors from Mumbai Indian actor-politicians Indian amateur radio operators Indian male film actors Hindi film producers Indian male singers Bollywood playback singers Indian television presenters Indian male voice actors Living people Best Actor National Film Award winners Male actors from Allahabad Recipients of the Padma Bhushan in arts Recipients of the Padma Shri in arts Who Wants to Be a Millionaire? India MPs 1984–1989 Lok Sabha members from Uttar Pradesh Recipients of the Padma Vibhushan in arts Indian male playback singers People named in the Panama Papers Film producers from Mumbai 20th-century Indian male actors 21st-century Indian male actors Male actors in Hindi television Indian male television actors Politicians from Allahabad Film producers from Uttar Pradesh People from New Alipore Filmfare Awards winners Filmfare Lifetime Achievement Award winners Screen Awards winners International Indian Film Academy Awards winners Zee Cine Awards winners Indian Hindus Dadasaheb Phalke Award recipients Actors from Mumbai People named in the Paradise Papers
1968
https://en.wikipedia.org/wiki/Apollo%2014
Apollo 14
Apollo 14 (January 31February 9, 1971) was the eighth crewed mission in the United States Apollo program, the third to land on the Moon, and the first to land in the lunar highlands. It was the last of the "H missions", landings at specific sites of scientific interest on the Moon for two-day stays with two lunar extravehicular activities (EVAs or moonwalks). The mission was originally scheduled for 1970, but was postponed because of the investigation following the failure of Apollo 13 to reach the Moon's surface, and the need for modifications to the spacecraft as a result. Commander Alan Shepard, Command Module Pilot Stuart Roosa, and Lunar Module Pilot Edgar Mitchell launched on their nine-day mission on Sunday, January 31, 1971, at 4:03:02 p.m. EST. En route to the lunar landing, the crew overcame malfunctions that might have resulted in a second consecutive aborted mission, and possibly, the premature end of the Apollo program. Shepard and Mitchell made their lunar landing on February 5 in the Fra Mauro formation – originally the target of Apollo 13. During the two walks on the surface, they collected of Moon rocks and deployed several scientific experiments. To the dismay of some geologists, Shepard and Mitchell did not reach the rim of Cone crater as had been planned, though they came close. In Apollo 14's most famous event, Shepard hit two golf balls he had brought with him with a makeshift club. While Shepard and Mitchell were on the surface, Roosa remained in lunar orbit aboard the Command and Service Module, performing scientific experiments and photographing the Moon, including the landing site of the future Apollo 16 mission. He took several hundred seeds on the mission, many of which were germinated on return, resulting in the so-called Moon trees, that were widely distributed in the following years. After liftoff from the lunar surface and a successful docking, the spacecraft was flown back to Earth where the three astronauts splashed down safely in the Pacific Ocean on February 9. Astronauts and key Mission Control personnel The mission commander of Apollo 14, Alan Shepard, one of the original Mercury Seven astronauts, became the first American to enter space with a suborbital flight on May 5, 1961. Thereafter, he was grounded by Ménière's disease, a disorder of the ear, and served as Chief Astronaut, the administrative head of the Astronaut Office. He had experimental surgery in 1968 which was successful and allowed his return to flight status. Shepard, at age 47, was the oldest U.S. astronaut to fly when he made his trip aboard Apollo 14, and he is the oldest person to walk on the Moon. Apollo 14's Command Module Pilot (CMP), Stuart Roosa, aged 37 when the mission flew, had been a smoke jumper before joining the Air Force in 1953. He became a fighter pilot and then in 1965 successfully completed Aerospace Research Pilot School (ARPS) at Edwards Air Force Base in California prior to his selection as a Group 5 astronaut the following year. He served as a capsule communicator (CAPCOM) for Apollo 9. The Lunar Module Pilot (LMP), Edgar Mitchell, aged 40 at the time of Apollo 14, joined the Navy in 1952 and served as a fighter pilot, beginning in 1954. He was assigned to squadrons aboard aircraft carriers before returning to the United States to further his education while in the Navy, also completing the ARPS prior to his selection as a Group 5 astronaut. He served on the support crew for Apollo 9 and was the LMP of the backup crew for Apollo 10. Shepard and his crew had originally been designated by Deke Slayton, Director of Flight Crew Operations and one of the Mercury Seven, as the crew for Apollo 13. NASA's management felt that Shepard needed more time for training given he had not flown in space since 1961, and chose him and his crew for Apollo 14 instead. The crew originally designated for Apollo 14, Jim Lovell as the commander, Ken Mattingly as CMP and Fred Haise as LMP, all of whom had backed up Apollo 11, was made the prime crew for Apollo 13 instead. Mitchell's commander on the Apollo 10 backup crew had been another of the original seven, Gordon Cooper, who had tentatively been scheduled to command Apollo 13, but according to author Andrew Chaikin, his casual attitude toward training resulted in him being not selected . Also on that crew, but excluded from further flights, was Donn Eisele, likely because of problems aboard Apollo 7, which he had flown, and because he had been involved in a messy divorce. Apollo 14's backup crew was Eugene A. Cernan as commander, Ronald E. Evans Jr. as CMP and Joe H. Engle as LMP. The backup crew, with Harrison Schmitt replacing Engle, would become the prime crew of Apollo 17. Schmitt flew instead of Engle because there was intense pressure on NASA to fly a scientist to the Moon (Schmitt was a geologist) and Apollo 17 was the last lunar flight. Engle, who had flown the X-15 to the edge of outer space, flew into space for NASA in 1981 on STS-2, the second Space Shuttle flight. During projects Mercury and Gemini, each mission had a prime and a backup crew. Apollo 9 commander James McDivitt believed meetings that required a member of the flight crew were being missed, so for Apollo a third crew of astronauts was added, known as the support crew. Usually low in seniority, support crew members assembled the mission's rules, flight plan, and checklists, and kept them updated; for Apollo 14, they were Philip K. Chapman, Bruce McCandless II, William R. Pogue and C. Gordon Fullerton. CAPCOMs, the individuals in Mission Control responsible for communications with the astronauts were Evans, McCandless, Fullerton and Haise. A veteran of Apollo 13, which had aborted before reaching the Moon, Haise put his training for that mission to use, especially during the EVAs, since both missions were targeted at the same place on the Moon. Had Haise walked on the Moon, he would have been the first Group 5 astronaut to do so, an honor that went to Mitchell. The flight directors during Apollo had a one-sentence job description, "The flight director may take any actions necessary for crew safety and mission success." For Apollo 14, they were: Pete Frank, Orange team; Glynn Lunney, Black team; Milt Windler, Maroon team and Gerry Griffin, Gold team. Preparation and training Prime and backup crews for both Apollo 13 and 14 were announced on August 6, 1969. Apollo 14 was scheduled for July 1970, but in January of that year, due to budget cuts that saw the cancellation of Apollo 20, NASA decided there would be two Apollo missions per year with 1970 to see Apollo 13 in April and Apollo 14 likely in October or November. The investigation into the accident which caused an abort of Apollo 13 delayed Apollo 14. On May 7, 1970, NASA Administrator Thomas O. Paine announced that Apollo 14 would launch no earlier than December 3, and the landing would be close to the site targeted by Apollo 13. The Apollo 14 astronauts continued their training. On June 30, 1970, following the release of the accident report and a NASA review of what changes to the spacecraft would be necessary, NASA announced that the launch would slip to no earlier than January 31, 1971. The crew of Apollo 14 trained together for 19 months after assignment to the mission, longer than any other Apollo crew to that point. In addition to the normal training workload, they had to supervise the changes to the command and service module (CSM) made as a result of the Apollo 13 investigation, much of which was delegated by Shepard to Roosa. Mitchell later stated, "We realized that if our mission failed—if we had to turn back—that was probably the end of the Apollo program. There was no way NASA could stand two failures in a row. We figured there was a heavy mantle on our shoulders to make sure we got it right." Before the abort of the Apollo 13 mission, the plan was to have Apollo 14 land near Littrow crater, in Mare Serenitatis, where there are features that were thought to be volcanic. After Apollo 13 returned, it was decided that its landing site, near Cone crater in the Fra Mauro formation, was scientifically more important than Littrow. The Fra Mauro formation is composed of ejecta from the impact event that formed Mare Imbrium, and scientists hoped for samples that originated deep under the Moon's surface. Cone crater was the result of a young, deep impact, and large enough to have torn through whatever debris was deposited since the Imbrium Event, which geologists hoped to be able to date. Landing at Fra Mauro would also allow orbital photography of another candidate landing site, the Descartes Highlands, which became the landing site for Apollo 16. Although Littrow went unvisited, a nearby area, Taurus-Littrow, was the landing site for Apollo 17. Apollo 14's landing site was located slightly closer to Cone crater than the point designated for Apollo 13. The change in landing site from Littrow to Fra Mauro affected the geological training for Apollo 14. Before the switch, the astronauts had been taken to volcanic sites on Earth; afterwards, they visited crater sites, such as the Ries Crater in West Germany and an artificial crater field created for astronaut training in Arizona's Verde Valley. The effectiveness of the training was limited by a lack of enthusiasm shown by Shepard, which set the tone for Mitchell. Harrison Schmitt suggested that the commander had other things on his mind, such as overcoming a ten-year absence from spaceflight and ensuring a successful mission after the near-disaster of Apollo 13. Roosa undertook training for his period alone in lunar orbit, when he would make observations of the Moon and take photographs. He had been impressed by the training given to Apollo 13 prime crew CMP Mattingly by geologist Farouk El-Baz and got El-Baz to agree to undertake his training. The two men pored over lunar maps depicting the areas the CSM would pass over. When Shepard and Mitchell were on their geology field trips, Roosa would be overhead in an airplane taking photographs of the site and making observations. El-Baz had Roosa make observations while flying his T-38 jet at a speed and altitude simulating the speed at which the lunar surface would pass below the CSM. Another issue that had marked Apollo 13 was the last-minute change of crew due to exposure to communicable disease. To prevent another such occurrence, for Apollo 14 NASA instituted what was called the Flight Crew Health Stabilization Program. Beginning 21 days before launch, the crew lived in quarters at the launch site, Florida's Kennedy Space Center (KSC), with their contacts limited to their spouses, the backup crew, mission technicians, and others directly involved in training. Those individuals were given physical examinations and immunizations, and crew movements were limited as much as possible at KSC and nearby areas. The Command and Service Modules were delivered to KSC on November 19, 1969; the ascent stage of the LM arrived on November 21 with the descent stage three days later. Thereafter, checkout, testing and equipment installation proceeded. The launch vehicle stack, with the spacecraft on top, was rolled out from the Vehicle Assembly Building to Pad 39A on November 9, 1970. Hardware Spacecraft The Apollo 14 spacecraft consisted of Command Module (CM) 110 and Service Module (SM) 110 (together CSM-110), called Kitty Hawk, and Lunar Module 8 (LM-8), called Antares. Roosa had chosen the CSM's call sign after the town in North Carolina where, in 1903, the Wright Brothers first flew their Wright Flyer airplane (also known as Kitty Hawk). Antares was named by Mitchell after the star in the constellation Scorpius that the astronauts in the LM would use to orient the craft for its lunar landing. Also considered part of the spacecraft were a Launch Escape System and a Spacecraft/Launch Vehicle Adapter, numbered SLA-17. The changes to the Apollo spacecraft between Apollo 13 and 14 were more numerous than with earlier missions, not only because of the problems with Apollo 13, but because of the more extensive lunar activities planned for Apollo 14. The Apollo 13 accident had been caused by the explosive failure of an oxygen tank, after the insulation of the internal wiring had been damaged by heating of the tank contents pre-launch—that the oxygen had gotten hot enough to damage the insulation had not been realized, since the protective thermostatic switches had failed because they were, through an error, not designed to handle the voltage applied during ground testing. The explosion damaged the other tank or its tubing, causing its contents to leak away. The changes in response included a redesign of the oxygen tanks, with the thermostats being upgraded to handle the proper voltage. A third tank was also added, placed in Bay1 of the SM, on the side opposite the other two, and was given a valve that could isolate it in an emergency, and allow it to feed the CM's environmental system only. The quantity probe in each tank was upgraded from aluminum to stainless steel. Also in response to the Apollo 13 accident, the electrical wiring in Bay4 (where the explosion had happened) was sheathed in stainless steel. The fuel cell oxygen supply valves were redesigned to isolate the Teflon-coated wiring from the oxygen. The spacecraft and Mission Control monitoring systems were modified to give more immediate and visible warnings of anomalies. The Apollo 13 astronauts had suffered shortages of water and of power after the accident. Accordingly, an emergency supply of of water was stored in Apollo 14's CM, and an emergency battery, identical to those that powered the LM's descent stage, was placed in the SM. The LM was modified to make the transfer of power from LM to CM easier. Other changes included the installation of anti-slosh baffles in the LM descent stage's propellant tanks. This would prevent the low fuel light from coming on prematurely, as had happened on Apollo 11 and 12. Structural changes were made to accommodate the equipment to be used on the lunar surface, including the Modular Equipment Transporter. Launch vehicle The Saturn V used for Apollo 14 was designated SA-509, and was similar to those used on Apollo 8 through 13. At , it was the heaviest vehicle yet flown by NASA, heavier than the launch vehicle for Apollo 13. A number of changes were made to avoid pogo oscillations, that had caused an early shutdown of the center J-2 engine on Apollo 13's S-II second stage. These included a helium gas accumulator installed in the liquid oxygen (LOX) line of the center engine, a backup cutoff device for that engine, and a simplified 2-position propellant utilization valve on each of the five J-2 engines. ALSEP and other lunar surface equipment The Apollo Lunar Surface Experiments Package (ALSEP) array of scientific instruments carried by Apollo 14 consisted of the Passive Seismic Experiment (PSE), Active Seismic Experiment (ASE), Suprathermal Ion Detector (SIDE), Cold Cathode Ion Gauge (CCIG), and Charged Particle Lunar Environmental Experiment (CPLEE). Two additional lunar surface experiments not part of the ALSEP were also flown, the Laser Ranging Retro-Reflector (LRRR or LR3), to be deployed in the ALSEP's vicinity, and the Lunar Portable Magnetometer (LPM), to be used by the astronauts during their second EVA. The PSE had been flown on Apollo 12 and 13, the ASE on Apollo 13, the SIDE on Apollo 12, the CCIG on Apollo 12 and 13, and the LRRR on Apollo 11. The LPM was new, but resembled equipment flown on Apollo 12. The ALSEP components flown on Apollo 13 were destroyed when its LM burned up in Earth's atmosphere. Deployment of the ALSEP, and of the other instruments, each formed one of Apollo 14's mission objectives. The PSE was a seismometer, similar to one left on the Moon by Apollo 12, and was to measure seismic activity in the Moon. The Apollo 14 instrument would be calibrated by the impact, after being jettisoned, of the LM's ascent stage, since an object of known mass and velocity would be impacting at a known location on the Moon. The Apollo 12 instrument would also be activated by the spent Apollo 14 S-IVB booster, which would impact the Moon after the mission entered lunar orbit. The two seismometers would, in combination with those left by later Apollo missions, constitute a network of such instruments at different locations on the Moon. The ASE would also measure seismic waves. It consisted of two parts. In the first, one of the crew members would deploy three geophones at distances up to from the ALSEP's Central Station, and on his way back from the furthest, fire thumpers every . The second consisted of four mortars (with their launch tubes), of different properties and set to impact at different distances from the experiment. It was hoped that the waves generated from the impacts would provide data about seismic wave transmission in the Moon's regolith. The mortar shells were not to be fired until the astronauts had returned to Earth, and in the event were never fired for fear they would damage other experiments. A similar experiment was successfully deployed, and the mortars launched, on Apollo 16. The LPM was to be carried during the second EVA and used to measure the Moon's magnetic field at various points. The SIDE measured ions on the lunar surface, including from the solar wind. It was combined with the CCIG, which was to measure the lunar atmosphere and detect if it varied over time. The CPLEE measured the particle energies of protons and electrons generated by the Sun that reached the lunar surface. The LRRR acts as a passive target for laser beams, allowing the measurement of the Earth/Moon distance and how it changes over time. The LRRRs from Apollo 11, 14 and 15 are the only experiments left on the Moon by the Apollo astronauts that are still returning data. Flown for the first time on Apollo 14 was the Buddy Secondary Life Support System (BSLSS), a set of flexible hoses that would enable Shepard and Mitchell to share cooling water should one of their Primary Life Support System (PLSS) backpacks fail. In such an emergency, the astronaut with the failed equipment would get oxygen from his Oxygen Purge System (OPS) backup cylinder, but the BSLSS would ensure he did not have to use oxygen for cooling, extending the life of the OPS. The OPSs used on Apollo 14 were modified from those used on previous missions in that the internal heaters were removed as unnecessary. Water bags were also taken to the lunar surface, dubbed "Gunga Dins", for insertion in the astronauts' helmets, allowing them sips of water during the EVAs. These had been flown on Apollo 13, but Shepard and Mitchell were the first to use them on the Moon. Similarly, Shepard was the first on the lunar surface to wear a spacesuit with commander's stripes: red stripes on arms, legs, and on the helmet, though one had been worn by Lovell on Apollo 13. These were instituted because of the difficulty in telling one spacesuited astronaut from the other in photographs. Modular Equipment Transporter The Modular Equipment Transporter (MET) was a two-wheeled handcart, used only on Apollo 14, intended to allow the astronauts to take tools and equipment with them, and store lunar samples, without needing to carry them. On later Apollo program missions, the self-propelled Lunar Roving Vehicle (LRV) was flown instead. The MET, when deployed for use on the lunar surface, was about long, wide and high. It had pressurized rubber tires wide and in diameter, containing nitrogen and inflated to about . The first use of tires on the Moon, these were developed by Goodyear and were dubbed their XLT (Experimental Lunar Tire) model. Fully loaded, the MET weighed about . Two legs combined with the wheels to provide four-point stability when at rest. Mission highlights Launch and flight to lunar orbit Apollo 14 launched from Launch Complex 39-A at KSC at 4:03:02 pm (21:03:02 UTC), January 31, 1971. This followed a launch delay due to weather of 40 minutes and 2 seconds; the first such delay in the Apollo program. The original planned time, 3:23 pm, was at the very start of the launch window of just under four hours; had Apollo 14 not launched during it, it could not have departed until March. Apollo 12 had launched during poor weather and twice been struck by lightning, as a result of which the rules had been tightened. Among those present to watch the launch were U.S. Vice President Spiro T. Agnew and the Prince of Spain, the future King Juan Carlos I. The mission would take a faster trajectory to the Moon than planned, and thus make up the time in flight. Because it had, just over two days after launch, the mission timers would be put ahead by 40 minutes and 3 seconds so that later events would take place at the times scheduled in the flight plan. After the vehicle reached orbit, the S-IVB third stage shut down, and the astronauts performed checks of the spacecraft before restarting the stage for translunar injection (TLI), the burn that placed the vehicle on course for the Moon. After TLI, the CSM separated from the S-IVB, and Roosa performed the transposition maneuver, turning it around in order to dock with the LM before the entire spacecraft separated from the stage. Roosa, who had practiced the maneuver many times, hoped to break the record for the least amount of propellant used in docking. But when he gently brought the modules together, the docking mechanism would not activate. He made several attempts over the next two hours, as mission controllers huddled and sent advice. If the LM could not be extracted from its place on the S-IVB, no lunar landing could take place, and with consecutive failures, the Apollo program might end. Mission Control proposed that they try it again with the docking probe retracted, hoping the contact would trigger the latches. This worked, and within an hour the joined spacecraft had separated from the S-IVB. The stage was set on a course to impact the Moon, which it did just over three days later, causing the Apollo 12 seismometer to register vibrations for over three hours. The crew settled in for its voyage to Fra Mauro. At 60:30 Ground Elapsed Time, Shepard and Mitchell entered the LM to check its systems; while there they photographed a wastewater dump from the CSM, part of a particle contamination study in preparation for Skylab. Two midcourse corrections were performed on the translunar coast, with one burn lasting 10.19 seconds and one lasting 0.65 seconds. Lunar orbit and descent At 81:56:40.70 into the mission (February 4 at 1:59:43 am EST; 06:59:43 UTC), the Service Propulsion System engine in the SM was fired for 370.84 seconds to send the craft into a lunar orbit with apocynthion of and pericynthion of . A second burn, at 86:10:52 mission time, sent the spacecraft into an orbit of by . This was done in preparation for the release of the LM Antares. Apollo 14 was the first mission on which the CSM propelled the LM to the lower orbit—though Apollo 13 would have done so had the abort not already occurred. This was done to increase the amount of hover time available to the astronauts, a safety factor since Apollo 14 was to land in rough terrain. After separating from the command module in lunar orbit, the LM Antares had two serious problems. First, the LM computer began getting an ABORT signal from a faulty switch. NASA believed the computer might be getting erroneous readings like this if a tiny ball of solder had shaken loose and was floating between the switch and the contact, closing the circuit. The immediate solution – tapping on the panel next to the switch – did work briefly, but the circuit soon closed again. If the problem recurred after the descent engine fired, the computer would think the signal was real and would initiate an auto-abort, causing the ascent stage to separate from the descent stage and climb back into orbit. NASA and the software teams at the Massachusetts Institute of Technology scrambled to find a solution. The software was hard-wired, preventing it from being updated from the ground. The fix made it appear to the system that an abort had already happened, and it would ignore incoming automated signals to abort. This would not prevent the astronauts from piloting the ship, though if an abort became necessary, they might have to initiate it manually. Mitchell entered the changes with minutes to go until planned ignition. A second problem occurred during the powered descent, when the LM landing radar failed to lock automatically onto the Moon's surface, depriving the navigation computer of vital information on the vehicle's altitude and vertical descent speed. After the astronauts cycled the landing radar breaker, the unit successfully acquired a signal near . Mission rules required an abort if the landing radar was out at , though Shepard might have tried to land without it. With the landing radar, Shepard steered the LM to a landing which was the closest to the intended target of the six missions that landed on the Moon. Lunar surface operations Shepard stated, after stepping onto the lunar surface, "And it's been a long way, but we're here." The first EVA began at 9:42 am EST (14:42 UTC) on February 5, 1971, having been delayed by a problem with the communications system which set back the start of the first EVA to five hours after landing. The astronauts devoted much of the first EVA to equipment offloading, deployment of the ALSEP and the US flag, as well as setting up and loading the MET. These activities were televised back to Earth, though the picture tended to degenerate during the latter portion of the EVA. Mitchell deployed the ASE's geophone lines, unreeling and emplacing the two lines leading out from the ALSEP's Central Station. He then fired the thumper explosives, vibrations from which would give scientists back on Earth information about the depth and composition of the lunar regolith. Of the 21 thumpers, five failed to fire. On the way back to the LM, the astronauts collected and documented lunar samples, and took photographs of the area. The first EVA lasted 4 hours, 47 minutes, 50 seconds. The astronauts had been surprised by the undulating ground, expecting flatter terrain in the area of the landing, and this became an issue on the second EVA, as they set out, MET in tow, for the rim of Cone crater. The craters that Shepard and Mitchell planned to use for navigational landmarks looked very different on the ground than on the maps they had, based on overhead shots taken from lunar orbit. Additionally, they consistently overestimated the distance they travelled. Mission Control and the CAPCOM, Fred Haise, could see nothing of this, as the television camera remained near the LM, but they worried as the clock ticked on the EVA, and monitored the heavy breathing and rapid heartbeats of the astronauts. They topped one ridge that they expected was the crater rim, only to view more such terrain beyond. Although Mitchell strongly suspected the rim was nearby, they had become physically exhausted from the effort. They were then instructed by Haise to sample where they were and then start moving back towards the LM. Later analysis using the pictures they took determined that they had come within about of the crater's rim. Images from the Lunar Reconnaissance Orbiter (LRO) show the tracks of the astronauts and the MET come to within 30 m of the rim. The difficulties faced by Shepard and Mitchell would emphasize the need for a means of transportation on the lunar surface with a navigation system, which was met by the Lunar Roving Vehicle, already planned to fly on Apollo 15. Once the astronauts returned to the vicinity of the LM and were again within view of the television camera, Shepard performed a stunt he had been planning for years in the event he reached the Moon, and which is probably what Apollo 14 is best remembered for. Shepard brought along a Wilson six iron golf club head, which he had modified to attach to the handle of the contingency sample tool, and two golf balls. Shepard took several one-handed swings (due to the limited flexibility of the EVA suit) and exuberantly exclaimed that the second ball went "miles and miles and miles" in the low lunar gravity. Mitchell then threw a lunar scoop handle as if it were a javelin. The "javelin" and one of the golf balls wound up in a crater together, with Mitchell's projectile a bit further. In an interview with Ottawa Golf, Shepard stated the other landed near the ALSEP. The second EVA lasted 4 hours, 34 minutes, 41 seconds. Shepard brought back the club, gave it to the USGA Museum in New Jersey, and had a replica made which he gave to the National Air and Space Museum. In February 2021, to commemorate Apollo 14's 50th anniversary, imaging specialist Andy Saunders, who had previously worked to produce the clearest image of Neil Armstrong on the Moon, produced new, digitally enhanced images that were used to estimate the final resting places of the two balls that Shepard hit - the first landed approximately 24 yards from the "tee", while the second managed 40 yards. Some geologists were pleased enough with the close approach to Cone crater to send a case of scotch to the astronauts while they were in post-mission quarantine, though their enthusiasm was tempered by the fact that Shepard and Mitchell had documented few of the samples they brought back, making it hard and sometimes impossible to discern where they came from. Others were less happy; Don Wilhelms wrote in his book on the geological aspects of Apollo, "the golf game did not set well with most geologists in light of the results at Cone crater. The total haul from the rim-flank of Cone ... was 16 Hasselblad photographs (out of a mission total of 417), six rock-size samples heavier than 50 g, and a grand total of 10 kg of samples, 9 kg of which are in one rock (sample 14321 [i.e., Big Bertha]). That is to say, apart from 14321 we have less than 1 kg of rock—962 g to be exact—from what in my opinion is the most important single point reached by astronauts on the Moon." Geologist Lee Silver stated, "The Apollo 14 crews did not have the right attitude, did not learn enough about their mission, had the burden of not having the best possible preflight photography, and they weren't ready." In their sourcebook on Apollo, Richard W. Orloff and David M. Harland doubted that if Apollo 13 had reached the Moon, Lovell, and Haise, given a more distant landing point, could have got as close to Cone crater as Shepard and Mitchell did. Lunar samples A total of of Moon rocks, or lunar samples, were brought back from Apollo 14. Most are breccias, which are rocks composed of fragments of other, older rocks. Breccias form when the heat and pressure of meteorite impacts fuse small rock fragments together. There were a few basalts that were collected in this mission in the form of clasts (fragments) in breccia. The Apollo 14 basalts are generally richer in aluminum and sometimes richer in potassium than other lunar basalts. Most lunar mare basalts collected during the Apollo program were formed from 3.0 to 3.8 billion years ago. The Apollo 14 basalts were formed 4.0 to 4.3 billion years ago, older than the volcanism known to have occurred at any of the mare locations reached during the Apollo program. In January 2019 research showed that Big Bertha, which weighs , has characteristics that make it likely to be a terrestrial (Earth) meteorite. Granite and quartz, which are commonly found on Earth but very rarely found on the Moon, were confirmed to exist on Big Bertha. To find the sample's age, the research team from Curtin University looked at bits of the mineral zircon embedded in its structure. "By determining the age of zircon found in the sample, we were able to pinpoint the age of the host rock at about four billion years old, making it similar to the oldest rocks on Earth," researcher Alexander Nemchin said, adding that "the chemistry of the zircon in this sample is very different from that of every other zircon grain ever analyzed in lunar samples, and remarkably similar to that of zircons found on Earth." This would mean Big Bertha is both the first discovered terrestrial meteorite and the oldest known Earth rock. Lunar orbit operations Roosa spent almost two days alone aboard Kitty Hawk, performing the first intensive program of scientific observation from lunar orbit, much of which was intended to have been done by Apollo 13. After Antares separated and its crew began preparations to land, Roosa in Kitty Hawk performed an SPS burn to send the CSM to an orbit of approximately , and later a plane change maneuver to compensate for the rotation of the Moon. Roosa took pictures from lunar orbit. The Lunar Topographic Camera, also known as the Hycon camera, was supposed to be used to image the surface, including the Descartes Highlands site being considered for Apollo 16, but it quickly developed a fault with the shutter that Roosa could not fix despite considerable help from Houston. Although about half of the photographic targets had to be scrubbed, Roosa was able to obtain photographs of Descartes with a Hasselblad camera and confirm that it was a suitable landing point. Roosa also used the Hasselblad to take photographs of the impact point of Apollo 13's S-IVB near Lansburg B crater. After the mission, troubleshooting found a tiny piece of aluminum contaminating the shutter control circuit, which caused the shutter to operate continuously. Roosa was able to see sunlight glinting off Antares and view its lengthy shadow on the lunar surface on Orbit 17; on Orbit 29 he could see the sun reflecting off the ALSEP. He also took astronomical photographs, of the Gegenschein, and of the Lagrangian point of the Sun-Earth system that lies beyond the Earth (L), testing the theory that the Gegenschein is generated by reflections off particles at L. Performing the bistatic radar experiment, he also focused Kitty Hawk VHF and S-band transmitters at the Moon so that they would bounce off and be detected on Earth in an effort to learn more about the depth of the lunar regolith. Return, splashdown and quarantine Antares lifted off from the Moon at 1:48:42 pm EST (18:48:42 UTC) on February 6, 1971. Following the first direct (first orbit) rendezvous on a lunar landing mission, docking took place an hour and 47 minutes later. Despite concerns based on the docking problems early in the mission, the docking was successful on the first attempt, though the LM's Abort Guidance System, used for navigation, failed just before the two craft docked. After crew, equipment, and lunar samples were transferred to Kitty Hawk, the ascent stage was jettisoned, and impacted the Moon, setting off waves registered by the seismometers from Apollo 12 and 14. A trans-earth injection burn took place on February 6 at 8:39:04 pm (February 7 at 01:39:04 UTC) taking 350.8 seconds, during Kitty Hawk 34th lunar revolution. During the trans-earth coast, two tests of the oxygen system were performed, one to ensure the system would operate properly with low densities of oxygen in the tanks, the second to operate the system at a high flow rate, as would be necessary for the in-flight EVAs scheduled for Apollo 15 and later. Additionally, a navigation exercise was done to simulate a return to Earth following a loss of communications. All were successful. During his rest periods on the voyage, Mitchell conducted ESP experiments without NASA's knowledge or sanction, attempting by prearrangement to send images of cards he had brought with him to four people on Earth. He stated after the mission that two of the four had gotten 51 out of 200 correct (the others were less successful), whereas random chance would have dictated 40. On the final evening in space, the crew conducted a press conference, with the questions submitted to NASA in advance and read to the astronauts by the CAPCOM. The command module Kitty Hawk splashed down in the South Pacific Ocean on February 9, 1971, at 21:05 [UTC], approximately south of American Samoa. After recovery by the ship USS New Orleans, the crew was flown to Pago Pago International Airport in Tafuna, then to Honolulu, then to Ellington Air Force Base near Houston in a plane containing a Mobile Quarantine Facility trailer before they continued their quarantine in the Lunar Receiving Laboratory. They remained there until their release from quarantine on February 27, 1971. The Apollo 14 astronauts were the last lunar explorers to be quarantined on their return from the Moon. They were the only Apollo crew to be quarantined both before and after the flight. Roosa, who worked in forestry in his youth, took several hundred tree seeds on the flight. These were germinated after the return to Earth, and were widely distributed around the world as commemorative Moon trees. Some seedlings were given to state forestry associations in 1975 and 1976 to mark the United States Bicentennial. Mission insignia The mission insignia is an oval depicting the Earth and the Moon, and an astronaut pin drawn with a comet trail. The pin is leaving Earth and is approaching the Moon. A gold band around the edge includes the mission and astronaut names. The designer was Jean Beaulieu, who based it on a sketch by Shepard, who had been head of the Astronaut Office and meant the pin to symbolize that through him, the entire corps was in spirit flying to the Moon. The backup crew spoofed the patch with its own version, with revised artwork showing a Wile E. Coyote cartoon character depicted as gray-bearded (for Shepard, who was 47 at the time of the mission and the oldest man on the Moon), pot-bellied (for Mitchell, who had a pudgy appearance) and red-furred (for Roosa's red hair), still on the way to the Moon, while Road Runner (for the backup crew) is already on the Moon, holding a U.S. flag and a flag labelled "1st Team". The flight name is replaced by "BEEP BEEP" and the backup crew's names are given. Several of these patches were hidden by the backup crew and found during the flight by the crew in notebooks and storage lockers in both the CSM Kitty Hawk and the LM Antares, and one patch was stored in the MET lunar handcart. One patch, attached to Shepard's PLSS, was worn on the lunar surface, and, mounted on a plaque, was presented by him to Cernan after the mission. Spacecraft locations The Apollo 14 command module Kitty Hawk is on display at the Apollo/Saturn V Center at the Kennedy Space Center Visitor Complex after being on display at the United States Astronaut Hall of Fame near Titusville, Florida, for several years. At the time of its transfer of ownership from NASA to the Smithsonian in July 1977, it was on display at the facilities of North American Rockwell (the company that had constructed it) in Downey, California. The SM reentered Earth's atmosphere and was destroyed, though there was no tracking or sightings of it. The S-IVB booster impacted the Moon on February4 at . The ascent stage of lunar module Antares impacted the Moon on February7, 1971, at 00:45:25.7 UT (February 6, 7:45 pm EST), at . Antares descent stage and the mission's other equipment remain at Fra Mauro at . Photographs taken in 2009 by the Lunar Reconnaissance Orbiter were released on July 17, and the Fra Mauro equipment was the most visible Apollo hardware at that time, owing to particularly good lighting conditions. In 2011, the LRO returned to the landing site at a lower altitude to take higher resolution photographs. Gallery See also Google Moon List of artificial objects on the Moon List of spacewalks and moonwalks 1965–1999 References Bibliography External links "Apollo 14" at Encyclopedia Astronautica Apollo 14 Traverse Map – United States Geological Survey (USGS) Apollo Mission Traverse Maps – Several maps showing routes of moonwalks Apollo 14 Science Experiments at the Lunar and Planetary InstituteNASA reports The Apollo Spacecraft: A Chronology NASA, NASA SP-4009 "Table 2-42. Apollo 14 Characteristics" from NASA Historical Data Book: Volume III: Programs and Projects 1969–1978 by Linda Neuman Ezell, NASA History Series (1988) "Masking the Abort Discrete" – by Paul Fjeld at the Apollo 14 Lunar Surface Journal. NASA. Detailed technical article describing the ABORT signal problem and its solution "Apollo 14 Technical Air-to-Ground Voice Transcription" (PDF) Manned Spacecraft Center, NASA, February 1971Multimedia' – slideshow by Life magazine "The Apollo Astronauts" – Interview with the Apollo 14 astronauts, March 31, 1971, from the Commonwealth Club of California Records at the Hoover Institution Archives "Apollo 14 Lunar Liftoff – Video" at Maniac World , with Cone crater 1971 in spaceflight 1971 in the United States Edgar Mitchell Stuart Roosa Alan Shepard Apollo program missions Articles containing video clips Extravehicular activity Crewed missions to the Moon Sample return missions Soft landings on the Moon Spacecraft which reentered in 1971 Spacecraft launched in 1971 January 1971 events February 1971 events Spacecraft launched by Saturn rockets 1971 on the Moon
2118
https://en.wikipedia.org/wiki/AVL%20tree
AVL tree
In computer science, an AVL tree (named after inventors Adelson-Velsky and Landis) is a self-balancing binary search tree. In an AVL tree, the heights of the two child subtrees of any node differ by at most one; if at any time they differ by more than one, rebalancing is done to restore this property. Lookup, insertion, and deletion all take time in both the average and worst cases, where is the number of nodes in the tree prior to the operation. Insertions and deletions may require the tree to be rebalanced by one or more tree rotations. The AVL tree is named after its two Soviet inventors, Georgy Adelson-Velsky and Evgenii Landis, who published it in their 1962 paper "An algorithm for the organization of information". It was the oldest self-balancing binary search tree data structure to be invented. AVL trees are often compared with red–black trees because both support the same set of operations and take time for the basic operations. For lookup-intensive applications, AVL trees are faster than red–black trees because they are more strictly balanced. Similar to red–black trees, AVL trees are height-balanced. Both are, in general, neither weight-balanced nor -balanced for any ; that is, sibling nodes can have hugely differing numbers of descendants. Definition Balance factor In a binary tree the balance factor of a node X is defined to be the height difference of its two child sub-trees. A binary tree is defined to be an AVL tree if the invariant holds for every node X in the tree. A node X with is called "left-heavy", one with is called "right-heavy", and one with is sometimes simply called "balanced". Properties Balance factors can be kept up-to-date by knowing the previous balance factors and the change in height – it is not necessary to know the absolute height. For holding the AVL balance information, two bits per node are sufficient. The height (counted as the maximal number of levels) of an AVL tree with nodes lies in the interval: where   is the golden ratio and This is because an AVL tree of height contains at least nodes where is the Fibonacci sequence with the seed values Operations Read-only operations of an AVL tree involve carrying out the same actions as would be carried out on an unbalanced binary search tree, but modifications have to observe and restore the height balance of the sub-trees. Searching Searching for a specific key in an AVL tree can be done the same way as that of any balanced or unbalanced binary search tree. In order for search to work effectively it has to employ a comparison function which establishes a total order (or at least a total preorder) on the set of keys. The number of comparisons required for successful search is limited by the height and for unsuccessful search is very close to , so both are in . Traversal As a read-only operation the traversal of an AVL tree functions the same way as on any other binary tree. Exploring all nodes of the tree visits each link exactly twice: one downward visit to enter the subtree rooted by that node, another visit upward to leave that node's subtree after having explored it. Once a node has been found in an AVL tree, the next or previous node can be accessed in amortized constant time. Some instances of exploring these "nearby" nodes require traversing up to links (particularly when navigating from the rightmost leaf of the root's left subtree to the root or from the root to the leftmost leaf of the root's right subtree; in the AVL tree of figure 1, navigating from node P to the next-to-the-right node Q takes 3 steps). Since there are links in any tree, the amortized cost is , or approximately 2. Insert When inserting a node into an AVL tree, you initially follow the same process as inserting into a Binary Search Tree. If the tree is empty, then the node is inserted as the root of the tree. If the tree is not empty, then we go down the root, and recursively go down the tree searching for the location to insert the new node. This traversal is guided by the comparison function. In this case, the node always replaces a NULL reference (left or right) of an external node in the tree i.e., the node is either made a left-child or a right-child of the external node. After this insertion, if a tree becomes unbalanced, only ancestors of the newly inserted node are unbalanced. This is because only those nodes have their sub-trees altered. So it is necessary to check each of the node's ancestors for consistency with the invariants of AVL trees: this is called "retracing". This is achieved by considering the balance factor of each node. Since with a single insertion the height of an AVL subtree cannot increase by more than one, the temporary balance factor of a node after an insertion will be in the range For each node checked, if the temporary balance factor remains in the range from –1 to +1 then only an update of the balance factor and no rotation is necessary. However, if the temporary balance factor is ±2, the subtree rooted at this node is AVL unbalanced, and a rotation is needed. With insertion as the code below shows, the adequate rotation immediately perfectly rebalances the tree. In figure 1, by inserting the new node Z as a child of node X the height of that subtree Z increases from 0 to 1. Invariant of the retracing loop for an insertion The height of the subtree rooted by Z has increased by 1. It is already in AVL shape. for (X = parent(Z); X != null; X = parent(Z)) { // Loop (possibly up to the root) // BF(X) has to be updated: if (Z == right_child(X)) { // The right subtree increases if (BF(X) > 0) { // X is right-heavy // ==> the temporary BF(X) == +2 // ==> rebalancing is required. G = parent(X); // Save parent of X around rotations if (BF(Z) < 0) // Right Left Case (see figure 3) N = rotate_RightLeft(X, Z); // Double rotation: Right(Z) then Left(X) else // Right Right Case (see figure 2) N = rotate_Left(X, Z); // Single rotation Left(X) // After rotation adapt parent link } else { if (BF(X) < 0) { BF(X) = 0; // Z’s height increase is absorbed at X. break; // Leave the loop } BF(X) = +1; Z = X; // Height(Z) increases by 1 continue; } } else { // Z == left_child(X): the left subtree increases if (BF(X) < 0) { // X is left-heavy // ==> the temporary BF(X) == -2 // ==> rebalancing is required. G = parent(X); // Save parent of X around rotations if (BF(Z) > 0) // Left Right Case N = rotate_LeftRight(X, Z); // Double rotation: Left(Z) then Right(X) else // Left Left Case N = rotate_Right(X, Z); // Single rotation Right(X) // After rotation adapt parent link } else { if (BF(X) > 0) { BF(X) = 0; // Z’s height increase is absorbed at X. break; // Leave the loop } BF(X) = -1; Z = X; // Height(Z) increases by 1 continue; } } // After a rotation adapt parent link: // N is the new root of the rotated subtree // Height does not change: Height(N) == old Height(X) parent(N) = G; if (G != null) { if (X == left_child(G)) left_child(G) = N; else right_child(G) = N; } else tree->root = N; // N is the new root of the total tree break; // There is no fall thru, only break; or continue; } // Unless loop is left via break, the height of the total tree increases by 1. In order to update the balance factors of all nodes, first observe that all nodes requiring correction lie from child to parent along the path of the inserted leaf. If the above procedure is applied to nodes along this path, starting from the leaf, then every node in the tree will again have a balance factor of −1, 0, or 1. The retracing can stop if the balance factor becomes 0 implying that the height of that subtree remains unchanged. If the balance factor becomes ±1 then the height of the subtree increases by one and the retracing needs to continue. If the balance factor temporarily becomes ±2, this has to be repaired by an appropriate rotation after which the subtree has the same height as before (and its root the balance factor 0). The time required is for lookup, plus a maximum of retracing levels ( on average) on the way back to the root, so the operation can be completed in time. Delete The preliminary steps for deleting a node are described in section Binary search tree#Deletion. There, the effective deletion of the subject node or the replacement node decreases the height of the corresponding child tree either from 1 to 0 or from 2 to 1, if that node had a child. Starting at this subtree, it is necessary to check each of the ancestors for consistency with the invariants of AVL trees. This is called "retracing". Since with a single deletion the height of an AVL subtree cannot decrease by more than one, the temporary balance factor of a node will be in the range from −2 to +2. If the balance factor remains in the range from −1 to +1 it can be adjusted in accord with the AVL rules. If it becomes ±2 then the subtree is unbalanced and needs to be rotated. (Unlike insertion where a rotation always balances the tree, after delete, there may be BF(Z) ≠ 0 (see figures 2 and 3), so that after the appropriate single or double rotation the height of the rebalanced subtree decreases by one meaning that the tree has to be rebalanced again on the next higher level.) The various cases of rotations are described in section Rebalancing. Invariant of the retracing loop for a deletion The height of the subtree rooted by N has decreased by 1. It is already in AVL shape. for (X = parent(N); X != null; X = G) { // Loop (possibly up to the root) G = parent(X); // Save parent of X around rotations // BF(X) has not yet been updated! if (N == left_child(X)) { // the left subtree decreases if (BF(X) > 0) { // X is right-heavy // ==> the temporary BF(X) == +2 // ==> rebalancing is required. Z = right_child(X); // Sibling of N (higher by 2) b = BF(Z); if (b < 0) // Right Left Case (see figure 3) N = rotate_RightLeft(X, Z); // Double rotation: Right(Z) then Left(X) else // Right Right Case (see figure 2) N = rotate_Left(X, Z); // Single rotation Left(X) // After rotation adapt parent link } else { if (BF(X) == 0) { BF(X) = +1; // N’s height decrease is absorbed at X. break; // Leave the loop } N = X; BF(N) = 0; // Height(N) decreases by 1 continue; } } else { // (N == right_child(X)): The right subtree decreases if (BF(X) < 0) { // X is left-heavy // ==> the temporary BF(X) == -2 // ==> rebalancing is required. Z = left_child(X); // Sibling of N (higher by 2) b = BF(Z); if (b > 0) // Left Right Case N = rotate_LeftRight(X, Z); // Double rotation: Left(Z) then Right(X) else // Left Left Case N = rotate_Right(X, Z); // Single rotation Right(X) // After rotation adapt parent link } else { if (BF(X) == 0) { BF(X) = -1; // N’s height decrease is absorbed at X. break; // Leave the loop } N = X; BF(N) = 0; // Height(N) decreases by 1 continue; } } // After a rotation adapt parent link: // N is the new root of the rotated subtree parent(N) = G; if (G != null) { if (X == left_child(G)) left_child(G) = N; else right_child(G) = N; } else tree->root = N; // N is the new root of the total tree if (b == 0) break; // Height does not change: Leave the loop // Height(N) decreases by 1 (== old Height(X)-1) } // If (b != 0) the height of the total tree decreases by 1. The retracing can stop if the balance factor becomes ±1 (it must have been 0) meaning that the height of that subtree remains unchanged. If the balance factor becomes 0 (it must have been ±1) then the height of the subtree decreases by one and the retracing needs to continue. If the balance factor temporarily becomes ±2, this has to be repaired by an appropriate rotation. It depends on the balance factor of the sibling Z (the higher child tree in figure 2) whether the height of the subtree decreases by one –and the retracing needs to continue– or does not change (if Z has the balance factor 0) and the whole tree is in AVL-shape. The time required is for lookup, plus a maximum of retracing levels ( on average) on the way back to the root, so the operation can be completed in time. Set operations and bulk operations In addition to the single-element insert, delete and lookup operations, several set operations have been defined on AVL trees: union, intersection and set difference. Then fast bulk operations on insertions or deletions can be implemented based on these set functions. These set operations rely on two helper operations, Split and Join. With the new operations, the implementation of AVL trees can be more efficient and highly-parallelizable. The function Join on two AVL trees and and a key will return a tree containing all elements in , as well as . It requires to be greater than all keys in and smaller than all keys in . If the two trees differ by height at most one, Join simply create a new node with left subtree , root and right subtree . Otherwise, suppose that is higher than for more than one (the other case is symmetric). Join follows the right spine of until a node which is balanced with . At this point a new node with left child , root and right child is created to replace c. The new node satisfies the AVL invariant, and its height is one greater than . The increase in height can increase the height of its ancestors, possibly invalidating the AVL invariant of those nodes. This can be fixed either with a double rotation if invalid at the parent or a single left rotation if invalid higher in the tree, in both cases restoring the height for any further ancestor nodes. Join will therefore require at most two rotations. The cost of this function is the difference of the heights between the two input trees. function JoinRightAVL(TL, k, TR) (l, k', c) = expose(TL) if (Height(c) <= Height(TR)+1) T' = Node(c, k, TR) if (Height(T') <= Height(l)+1) then return Node(l, k', T') else return rotateLeft(Node(l, k', rotateRight(T'))) else T' = JoinRightAVL(c, k, TR) T'' = Node(l, k', T') if (Height(T') <= Height(l)+1) return T'' else return rotateLeft(T'') function JoinLeftAVL(TL, k, TR) /* symmetric to JoinRightAVL */ function Join(TL, k, TR) if (Height(TL)>Height(TR)+1) return JoinRightAVL(TL, k, TR) if (Height(TR)>Height(TL)+1) return JoinLeftAVL(TL, k, TR) return Node(TL, k, TR) Here Height(v) is the height of a subtree (node) . (l,k,r) = expose(v) extracts 's left child , the key of 's root, and the right child . Node(l,k,r) means to create a node of left child , key , and right child . To split an AVL tree into two smaller trees, those smaller than key , and those larger than key , first draw a path from the root by inserting into the AVL. After this insertion, all values less than will be found on the left of the path, and all values greater than will be found on the right. By applying Join, all the subtrees on the left side are merged bottom-up using keys on the path as intermediate nodes from bottom to top to form the left tree, and the right part is asymmetric. The cost of Split is , order of the height of the tree. function Split(T, k) if (T = nil) return (nil, false, nil) (L,m,R) = expose(T) if (k = m) return (L, true, R) if (k<m) (L',b,R') = Split(L,k) return (L', b, Join(R', m, R)) if (k>m) (L',b,R') = Split(R, k) return (Join(L, m, L'), b, R')) The union of two AVL trees and representing sets and , is an AVL that represents . function Union(t1, t2): if t1 = nil: return t2 if t2 = nil: return t1 (t<, b, t>) = Split(t2, t1.root) return Join(Union(left(t1), t<), t1.root, Union(right(t1), t>)) Here, Split is presumed to return two trees: one holding the keys less its input key, one holding the greater keys. (The algorithm is non-destructive, but an in-place destructive version exists as well.) The algorithm for intersection or difference is similar, but requires the Join2 helper routine that is the same as Join but without the middle key. Based on the new functions for union, intersection or difference, either one key or multiple keys can be inserted to or deleted from the AVL tree. Since Split calls Join but does not deal with the balancing criteria of AVL trees directly, such an implementation is usually called the "join-based" implementation. The complexity of each of union, intersection and difference is for AVL trees of sizes and . More importantly, since the recursive calls to union, intersection or difference are independent of each other, they can be executed in parallel with a parallel depth . When , the join-based implementation has the same computational DAG as single-element insertion and deletion. Rebalancing If during a modifying operation the height difference between two child subtrees changes, this may, as long as it is < 2, be reflected by an adaption of the balance information at the parent. During insert and delete operations a (temporary) height difference of 2 may arise, which means that the parent subtree has to be "rebalanced". The given repair tools are the so-called tree rotations, because they move the keys only "vertically", so that the ("horizontal") in-order sequence of the keys is fully preserved (which is essential for a binary-search tree). Let X be the node that has a (temporary) balance factor of −2 or +2. Its left or right subtree was modified. Let Z be the higher child (see figures 2 and 3). Note that both children are in AVL shape by induction hypothesis. In case of insertion this insertion has happened to one of Z's children in a way that Z's height has increased. In case of deletion this deletion has happened to the sibling t1 of Z in a way so that t1's height being already lower has decreased. (This is the only case where Z's balance factor may also be 0.) There are four possible variants of the violation: And the rebalancing is performed differently: Thereby, the situations are denoted as where C (= child direction) and B (= balance) come from the set } with The balance violation of case is repaired by a simple rotation whereas the case is repaired by a double rotation The cost of a rotation, either simple or double, is constant. Simple rotation Figure 2 shows a Right Right situation. In its upper half, node X has two child trees with a balance factor of +2. Moreover, the inner child t23 of Z (i.e., left child when Z is right child, or right child when Z is left child) is not higher than its sibling t4. This can happen by a height increase of subtree t4 or by a height decrease of subtree t1. In the latter case, also the pale situation where t23 has the same height as t4 may occur. The result of the left rotation is shown in the lower half of the figure. Three links (thick edges in figure 2) and two balance factors are to be updated. As the figure shows, before an insertion, the leaf layer was at level h+1, temporarily at level h+2 and after the rotation again at level h+1. In case of a deletion, the leaf layer was at level h+2, where it is again, when t23 and t4 were of same height. Otherwise the leaf layer reaches level h+1, so that the height of the rotated tree decreases. Code snippet of a simple left rotation node *rotate_Left(node *X, node *Z) { // Z is by 2 higher than its sibling t23 = left_child(Z); // Inner child of Z right_child(X) = t23; if (t23 != null) parent(t23) = X; left_child(Z) = X; parent(X) = Z; // 1st case, BF(Z) == 0, // only happens with deletion, not insertion: if (BF(Z) == 0) { // t23 has been of same height as t4 BF(X) = +1; // t23 now higher BF(Z) = –1; // t4 now lower than X } else { // 2nd case happens with insertion or deletion: BF(X) = 0; BF(Z) = 0; } return Z; // return new root of rotated subtree } Double rotation Figure 3 shows a Right Left situation. In its upper third, node X has two child trees with a balance factor of +2. But unlike figure 2, the inner child Y of Z is higher than its sibling t4. This can happen by the insertion of Y itself or a height increase of one of its subtrees t2 or t3 (with the consequence that they are of different height) or by a height decrease of subtree t1. In the latter case, it may also occur that t2 and t3 are of the same height. The result of the first, the right, rotation is shown in the middle third of the figure. (With respect to the balance factors, this rotation is not of the same kind as the other AVL single rotations, because the height difference between Y and t4 is only 1.) The result of the final left rotation is shown in the lower third of the figure. Five links (thick edges in figure 3) and three balance factors are to be updated. As the figure shows, before an insertion, the leaf layer was at level h+1, temporarily at level h+2 and after the double rotation again at level h+1. In case of a deletion, the leaf layer was at level h+2 and after the double rotation it is at level h+1, so that the height of the rotated tree decreases. Code snippet of a right-left double rotation node *rotate_RightLeft(node *X, node *Z) { // Z is by 2 higher than its sibling Y = left_child(Z); // Inner child of Z // Y is by 1 higher than sibling t3 = right_child(Y); left_child(Z) = t3; if (t3 != null) parent(t3) = Z; right_child(Y) = Z; parent(Z) = Y; t2 = left_child(Y); right_child(X) = t2; if (t2 != null) parent(t2) = X; left_child(Y) = X; parent(X) = Y; // 1st case, BF(Y) == 0, // only happens with deletion, not insertion: if (BF(Y) == 0) { BF(X) = 0; BF(Z) = 0; } else // other cases happen with insertion or deletion: if (BF(Y) > 0) { // t3 was higher BF(X) = –1; // t1 now higher BF(Z) = 0; } else { // t2 was higher BF(X) = 0; BF(Z) = +1; // t4 now higher } BF(Y) = 0; return Y; // return new root of rotated subtree } Comparison to other structures Both AVL trees and red–black (RB) trees are self-balancing binary search trees and they are related mathematically. Indeed, every AVL tree can be colored red–black, but there are RB trees which are not AVL balanced. For maintaining the AVL (or RB tree's invariants), rotations play an important role. In the worst case, even without rotations, AVL or RB insertions or deletions require inspections and/or updates to AVL balance factors (or RB colors). RB insertions and deletions and AVL insertions require from zero to three tail-recursive rotations and run in amortized time, thus equally constant on average. AVL deletions requiring rotations in the worst case are also on average. RB trees require storing one bit of information (the color) in each node, while AVL trees mostly use two bits for the balance factor, although, when stored at the children, one bit with meaning «lower than sibling» suffices. The bigger difference between the two data structures is their height limit. For a tree of size an AVL tree's height is at most where   the golden ratio,   and  . a RB tree's height is at most  . AVL trees are more rigidly balanced than RB trees with an asymptotic relation AVL/RB ≈0.720 of the maximal heights. For insertions and deletions, Ben Pfaff shows in 79 measurements a relation of AVL/RB between 0.677 and 1.077 with median ≈0.947 and geometric mean ≈0.910. See also WAVL tree Splay tree Scapegoat tree B-tree T-tree List of data structures References Further reading Donald Knuth. The Art of Computer Programming, Volume 3: Sorting and Searching, Third Edition. Addison-Wesley, 1997. . Pages 458–475 of section 6.2.3: Balanced Trees. . External links 1962 in computing Articles with example pseudocode Binary trees Soviet inventions Search trees
2286
https://en.wikipedia.org/wiki/Tank%20destroyer
Tank destroyer
A tank destroyer, tank hunter, tank killer, or self-propelled anti-tank gun is a type of armoured fighting vehicle, armed with a direct fire artillery gun or missile launcher, designed specifically to engage and destroy enemy tanks, often with limited operational capacities. While tanks are designed for front-line combat, combining operational mobility and tactical offensive and defensive capabilities and performing all primary tasks of the armoured troops, the tank destroyer is specifically designed to take on enemy tanks and other armoured fighting vehicles. Many are based on a tracked tank chassis, while others are wheeled. Since World War II, gun-armed powerful tank destroyers have fallen out of favor as armies have favored multirole main battle tanks. However, lightly armoured anti-tank guided missile (ATGM) carriers are commonly used for supplementary long-range anti-tank work. The resurgence of expeditionary warfare in the first two decades of the 21st century has seen the emergence of gun-armed wheeled vehicles, sometimes called "protected gun systems", which may bear a superficial resemblance to tank destroyers, but are employed as direct fire support units typically providing support in low-intensity operations such as the wars in Iraq and Afghanistan. World War II Dedicated anti-tank vehicles made their first major appearance in the Second World War as combatants developed effective armoured vehicles and tactics. Some were little more than stopgap solutions, mounting an anti-tank gun on a tracked vehicle to give mobility, while others were more sophisticated designs. An example of the development of tank destroyer technology throughout the war are the Marder III and Jagdpanzer 38 vehicle, that were very different in spite of being based on the same chassis: Marder was straightforwardly an anti-tank gun on tracks whereas the Jagdpanzer 38 traded some firepower (its 7.5 cm Pak 39, designed to operate within the confines of a fully armoured fighting compartment, fires the same projectiles from a reduced propellant charge compared to Marder's 7.5 cm Pak 40) for better armour protection and ease of concealment on the battlefield. Except for most American designs, all tank destroyers were turretless vehicles with fixed or casemate superstructures. When a tank destroyer was used against enemy tanks from a defensive position such as by ambush, the lack of a rotating turret was not particularly critical, while the lower silhouette was highly desirable. The turretless design allowed accommodation of a more powerful gun, typically a dedicated anti-tank gun (in lieu of a regular tank's general-purpose main gun that fired both anti-tank and high explosive ammunition) that had a longer barrel than could be mounted in a turreted tank on the same chassis. The lack of a turret increased the vehicle's internal volume, allowing for increased ammunition stowage and crew comfort. Eliminating the turret let the vehicle carry thicker armour, and also let this armour be concentrated in the hull. Sometimes there was no armoured roof (only a weather cover) to keep the overall weight down to the limit that the chassis could bear. The absence of a turret meant that tank destroyers could be manufactured significantly cheaper, faster, and more easily than the tanks on which they were based, and they found particular favor when production resources were lacking. Germany The first German tank destroyers were the Panzerjäger ("tank hunters"), which mounted an existing anti-tank gun on a convenient chassis for mobility, usually with just a three-sided gun shield for crew protection. For instance, 202 obsolete Panzer I light tanks were modified by removing the turret and were rebuilt as the Panzerjäger I self-propelled 4.7 cm PaK(t). Similarly, Panzer II tanks were used on the eastern front. Captured Soviet anti-tank guns were mounted on modified Panzer II chassis, producing the Marder II self-propelled anti-tank gun. The most common mounting was a German anti-tank gun on the Czech Panzer 38(t) chassis as the Marder III. The Panzer 38(t) chassis was also used to make the Jagdpanzer 38 casemate style tank destroyer. The Panzerjäger series continued up to the equipped Nashorn. German tank destroyers based on the Panzer III medium tank and later German tanks had more armour than their tank counterparts. One of the more successful German tank destroyers was designed as a self-propelled artillery gun, the Sturmgeschütz III. Based on the Panzer III tank chassis, the Sturmgeschütz III was originally fitted with a short barreled low-velocity howitzer-like gun, and was assigned to the artillery arm for infantry fire support as an assault gun. Later, after encountering Soviet tanks, it was refitted with a comparatively short-barreled high-velocity anti-tank gun, usually with a muzzle brake, enabling it to function as a tank destroyer. The Sturmgeschütz III from its 1938 origin used a new casemate-style superstructure with an integrated design, similar to the later Jagdpanzer vehicle designs' superstructure, to completely enclose the crew. It was employed in infantry support and offensive armoured operations as well as in the defensive anti-tank role. The StuG III assault gun was Germany's most-produced fully tracked armoured fighting vehicle during World War II, and second-most produced German armoured combat vehicle of any type after the Sd.Kfz. 251 half-track. Although the early German Panzerjäger carried more effective weapons than the tanks on which they were based, they were generally lacking in protection for the crew, having thinly armoured open-topped superstructures. The "open-topped" design format of the Panzerjäger vehicles was succeeded by the Jagdpanzer ("hunting tanks"), which mounted the gun in true casemate-style superstructures, completely enclosing the crew compartment in armor that was usually integral to the hull. The first of these Jagdpanzers was the 70-ton Ferdinand (later renamed Elefant), based on the chassis, hulls, and drive systems of ninety-one Porsche VK4501 (P) heavy tanks, mounting a long-barreled 88 mm cannon in an added casemate, more like the earlier Panzerjägers had with their added-on armour shielding for the gun crew, but in the Ferdinand completely enclosing the gun and firing crew in the added casemate, as the later purpose-built Jagdpanzers would. However, the Ferdinand was mechanically unreliable and difficult to maneuver, and once all ninety-one unturreted "Porsche Tiger" hulls/drive systems were converted, no more were built. The German Army had more success with the Jagdpanther. Introduced in mid-1944, the Jagdpanther, of which some 415 examples were produced, was considered the best of the casemate-design Jagdpanzer designs. It featured the same powerful PaK 43 88 mm cannon used on the unwieldy Elefant, now fitted to the chassis of the medium Panther tank, providing greatly improved armour-penetrating capability in a medium-weight vehicle. Facing an increasingly defensive war, the German Army turned to larger and more powerfully armed Jagdpanzer designs, and in July 1944 the first Jagdtiger rolled off the production line; it was the heaviest German armoured fighting vehicle to go into active service. The Jagdtiger was based on the Tiger II heavy tank featured a very large 128 mm PaK 44 cannon and heavy armour protection. Only 88 Jagdtiger vehicles were produced, barely matching the total number of the earlier Ferdinand / Elefant vehicles. They were first deployed to combat units in September 1944. The decision of German armoured vehicle designers to use a casemate-style superstructure for all tank destroyers had the advantage of a reduced silhouette, allowing the crew to more frequently fire from defilade ambush positions. Such designs were also easier and faster to manufacture and offered good crew protection from artillery fire and shell splinters. However, the lack of a rotating turret limited the gun's traverse to a few degrees. This meant that the driver normally had to turn the entire tank onto its target, a much slower process than simply rotating a powered turret. If the vehicle became immobilized due to engine failure or track damage, it could not rotate its gun to counter opposing tanks, making it highly vulnerable to counterfire. This vulnerability was later exploited by opposing tank forces. Even the largest and most powerful of German tank destroyers were found abandoned on the field after a battle, having been immobilized by one or more hits by high explosive (HE) or armour-piercing (AP) shells to the track or front drive sprocket. Italy The most famous Italian tank destroyer of the Second World War was a self-propelled gun. The Semovente da 75/18, based on the M13/40 frame, was developed to support front-line infantry, and therefore had fixed armament: a 75 mm gun in casemate. However, thanks to its low height (185 cm) and the caliber of its gun the 75/18 also had good results in anti-tank combat, fighting against British and American (but not Soviet) units. After the Armistice of 1943, the 75/18 remained in use by German forces. Built on the same frame, the Semovente da 105/25 was equipped with a 105 mm gun and known as "bassotto" (Italian for dachshund) due to its lower height. As manufacturing began in 1943, the 105/25 was used by German forces. A further development was the Semovente da 75/46, which had a longer gun than the 75/18 and inclined armour 100 mm thick, making it similar to Sturmgeschütz III. Only 11 of these were manufactured. Before the Semovente da 75/18, the L40, built on an L6/40 light tank chassis, saw action in Africa and in Russia, but with disappointing results. Japan The Type 1 Ho-Ni I was the first self-propelled gun design of the Imperial Japanese Army. They were meant to be self-propelled artillery and tank destroyers for armoured divisions. The plan was for the Type 1 Ho-Ni I gun tank to form part of a fire support company in each of the tank regiments. The Type 1 Ho-Ni I was developed by using the existing Type 97 Chi-Ha medium tank chassis and engine, and replacing the gun turret with a Type 90 75 mm field gun mounted in an open casemate with frontal and side armour only. They entered service in 1942 and were first deployed in combat at the Battle of Luzon in the Philippines in 1945. Some were used in static entrenched positions. A variant, known as the Type 1 Ho-Ni II mounted a Type 91 105 mm howitzer and had a slightly changed superstructure as far as the side armor with re-positioned observation visors. Production began in 1943, with only 54 completed. The other variant produced was the Type 3 Ho-Ni III, which mounted a Type 3 75 mm tank gun in a completely enclosed armored casemate to address the issue of crew protection in close combat. The welded superstructure had sloped armour and the gun mount had additional stamped armour plate. The total number produced of all three types in the Ho-Ni series were 111 units. Most of the Ho-Ni units were retained within the Japanese home islands to form part of the defenses against the projected American invasion, and did not see combat before the surrender of Japan. The Type 2 Ho-I Gun tank used the Type 1 Chi-He medium tank chassis. It was designed as a self-propelled howitzer, mounting a short barreled Type 99 75 mm gun to provide close-in fire support. For deployment, the gun tank was intended to be used in a fire support company for each of the tank regiments. No Type 2 Ho-I gun tanks are known to have engaged in combat prior to Japan's surrender. The prototype was built in 1942 and 31 units were produced in 1944. The Type 4 Ho-Ro self-propelled artillery used a modified Type 97 chassis. On to this platform, a Type 38 150 mm howitzer was mounted. The main gun could fire Type 88 APHE rounds and HEAT rounds. Given its breech loader, the maximum rate of fire was only 5 rounds per minute. The gun's elevation was restricted to 30 degrees by the construction of the chassis. Other design issues included the fact that although the gun crew was protected by a gun shield with armour thickness of 25 mm at the front, the shield only extended a very short distance on the sides; leaving the rest of the sides and back exposed. They were rushed into service, deployed and saw combat during the Philippines Campaign in the last year of World War II. Remaining units were deployed to Okinawa in ones and twos for island defense during the Battle of Okinawa, but were severely outnumbered by American artillery. Soviet Union As with the Germans of 1943, most of the Soviet designs mounted anti-tank guns, with limited traverse in casemate-style turretless hulls, in a general design format looking much like the Germans' own Jagdpanzer vehicles. The results were smaller, lighter, and simpler to build weapons that could carry larger guns than any contemporary tank, including the King Tiger. The Soviets produced high numbers of the SU-85 and SU-100 self-propelled guns based on the same chassis as the T-34 medium tank; the heavier-duty powertrain and hull of the IS-2 heavy tank were instead used to produce the heavier-hitting -armed ISU-122 and -armed ISU-152, both of which had impressive anti-tank capabilities earning each of them the Russian nickname Zveroboy ("beast killer") for their ability to destroy German Tigers, Panthers and Elefants. The predecessor of the ISU 152 was the SU-152, built on the KV-1s chassis and shared many similarities (including its gun) with the ISU-152. The ISU-152 built as a heavy assault gun, relied on the weight of the shell fired from its M-1937/43 howitzer to defeat tanks. In 1943, the Soviets also shifted all production of light tanks like the T-70 to much simpler and better-armed SU-76 self-propelled guns, which used the same drive train. The SU-76 was originally designed as an anti-tank vehicle, but was soon relegated to the infantry-support role. United States U.S. Army and counterpart British designs were very different in conception. U.S. doctrine was based, in light of the fall of France, on the perceived need to defeat German blitzkrieg tactics, and U.S. units expected to face large numbers of German tanks, attacking on relatively narrow fronts. These were expected to break through a thin screen of anti-tank guns, hence the decision that the main anti-tank units—the Tank Destroyer (TD) battalions—should be concentrated and very mobile. In practice, such German attacks rarely happened. Throughout the war, only one battalion ever fought in an engagement like that originally envisaged (the 601st, at the Battle of El Guettar). The Tank Destroyer Command eventually numbered over 100,000 men and 80 battalions each equipped with 36 self-propelled tank destroyers or towed guns. Only a few shots were expected to be fired from any firing position. Strong reconnaissance elements were provided so that TDs could use pre-arranged firing positions to best advantage. Flanking fire by TDs was emphasized, both to penetrate thinner enemy side armour, and to reduce the likelihood of accurate enemy return fire. All American tank destroyers were officially known by exactly the same collective term used for American self-propelled artillery ordnance, "gun motor carriage". The designs were intended to be very mobile and heavily armed. Most of the tank-hull based designs used special open-topped turrets of a differing design from the original tank it was based on, which was meant to both save weight and to accommodate a larger gun. The earliest expedient design was an M3 Half-track mounting an 75 mm Field Gun M1897 on M2 Carriage in a limited-traverse mount, and called the 75 mm Gun Motor Carriage M3. Another, considerably less successful, early design mounted a 37-mm anti-tank gun in the bed of a Dodge 3/4-ton truck—the 37-mm GMC M6. By far the most common US design, and the first that was fully tracked and turreted (which became the American hallmark of World War II "tank destroyer" design) was the 3in Gun Motor Carriage M10, later supplemented by the 90 mm Gun Motor Carriage M36—both based on the M4 Sherman hull and powertrain—and the 76 mm Gun Motor Carriage M18 (Hellcat), based on a unique hull and powertrain design, with a slight visual resemblance to what was used for the later M24 Chaffee light tank. The M18 came closest to the US ideal; the vehicle was very fast, small, and mounted a gun in a roofless open turret. The M36 Jackson GMC possessed the only American-origin operational gun that could rival the vaunted 88 mm German anti-tank ordnance, the 90 mm M3 gun, and the M36 remained in service well after World War II. The only dedicated American casemate hull design fighting vehicle of any type built during the war, that resembled the German and Soviet tank destroyers in hull and general gun mounting design, was the experimental T28 Super Heavy Tank, which mounted a 105 mm T5E1 long-barrel cannon, which had a maximum firing range of 12 miles (20 km), and was originally designed as a self-propelled assault gun to breach Germany's Siegfried Line defenses. Of these tank destroyers, only the gun of the M36 proved effective against the frontal armour of Germans' larger armored vehicles at long range. The open top and light armour made these tank destroyers vulnerable to anything greater than small-arms fire. As the number of German tanks encountered by American forces steadily decreased throughout the war, most battalions were split up and assigned to infantry units as supporting arms, fighting as assault guns or being used essentially as tanks. In this sense they were an alternative to the Independent tank battalions that were attached to various Infantry Divisions. The expectation that German tanks would be engaged in mass formation was a failed assumption. In reality, German attacks effectively used combined arms on the ground, fighting cohesively. American tank destroyer battalions comprised three tank destroyer companies supported by nine security sections. The single-purpose tactics of the tank destroyer battalion failed to account for non-tank threats. In the 1950s the goal of providing airborne forces with a parachute-capable self-propelled anti-tank weapon led to the deployment of the M56 Scorpion and M50 Ontos. The concept later led to the M551 Sheridan light tank of the mid-1960s. United Kingdom British tanks in the early years of the war, both infantry tanks and cruiser tanks, were (with the exception of the pre-war Matilda I design) equipped with a gun capable of use against contemporary enemy tanks—the 40 mm Ordnance QF 2 pounder. This was replaced with the 57 mm Ordnance QF 6 pounder when that became available. There was extra impetus given to the development of anti-tank weaponry, which culminated in the 76mm Ordnance QF 17 pounder, widely considered one of the best anti-tank guns of the war. Towed anti-tank guns were the domain of the Royal Artillery and vehicles adapted to mount artillery, including anti-tank self-propelled guns such as the Deacon (6pdr on an armoured wheeled truck chassis) and Archer (17pdr on tracked chassis) and US-supplied vehicles, were their preserve rather than the Royal Armoured Corps. The self-propelled guns that were built in the "tank destroyer" mould came about through the desire to field the QF 17 pounder anti-tank gun and simultaneous lack of suitable standard tanks to carry it. As a result, they were of a somewhat extemporized nature. Mounting the gun on the Valentine tank chassis in a fixed superstructure gave the Archer, looking somewhat like the light-chassis German Marder III in appearance. The 17 pounder was also used to re-equip the US-supplied M10 Tank Destroyer, replacing the American 3-inch gun to produce the 17pdr SP Achilles. In 1942 the General Staff agreed on investigating self-propelled mountings of the 6-pounder, 17-pounder, 3-inch 20cwt guns and the 25-pounder field gun/howitzer on the Matilda II, Valentine, Crusader and Cavalier (Cruiser Mark VII) tank chassis. In October 1942 it was decided to progress using the Valentine chassis with a 17-pdr (which would become Archer) and 25-pdr (which entered service as Bishop). While there was a general move to a general purpose gun that was usable against both tanks and in supporting infantry, there was a need to put the 17 pdr into a tank for use against the enemy's heavy tanks. The Cruiser Mk VIII Challenger was a project to bring a 17 pdr tank into use to support the Cromwell cruiser tank. Delays led to it being outnumbered in use by the Sherman Firefly—but a derivative of Challenger was the more or less open-topped variant Avenger, which was delayed until post war before entering service. A cut-down 17 pdr, the 77mmHV was used to equip the Comet tank in the last year of the war. The closest the British came to developing an armoured tank destroyer in the vein of the German Jagdpanzers or Soviet ISU series was the Churchill 3-inch Gun Carrier—a Churchill tank chassis with a boxy superstructure in place of the turret and mounting a 3-inch anti-aircraft gun. Although a number were ordered and fifty delivered in 1942, they were not put into service as the immediate threat passed. The design was rejected in favor of developing a 17 pounder armed Cromwell tank variant, ultimately leading to the Comet tank. The Tortoise "heavy assault tank", intended for use in breaking through fixed defensive lines, was well armoured and had a very powerful 32-pounder (94 mm) gun, but did not reach service use. By 1944, a number of the Shermans in British use were being converted to Sherman Fireflies by adding the QF 17 pounder gun. Initially this gave each troop (platoon) of Shermans one powerfully armed tank. By war's end—through the production of more Fireflies and the replacement of Shermans by British tanks—about 50% of Shermans in British service were Fireflies. The Sherman Firefly, however, is not considered a tank destroyer since it could still perform the other duties of the regular M4 Sherman, albeit the Firefly was less capable due to the late development of a HE round for the QF 17 pounder. Romania Until 1942, the Romanian tank force was equipped exclusively with obsolete R-1, R-2 and R35 tanks. Having faced big problems against Soviet T-34 and KV-1 tanks on the Eastern Front, the Romanian Army leadership sought for ways to improve its anti-tank capabilities. The initial plan was the creation of a tank comparable in characteristics to the T-34; instead, Romania went for a number of tank destroyers, since they were more adequate for its industry. The Mareșal is probably the best known Romanian AFV from the war; historians Steven Zaloga and Mark Axworthy state that it inspired the design of the later German Hetzer. Standing at only around 1.5 m tall, which would have made it very difficult to hit for its enemies, the Mareșal was a lightly armored, but highly mobile vehicle. It was armed with the Romanian 75 mm Reșița M1943 anti-tank gun, which proved to be among the best of its class during World War II, according to Mark Axworthy. During tests, the Mareșal proved to be superior in many aspects to the StuG III G, against which it competed. Those facts suggest that the Mareșal would have been an effective tank destroyer, had it been deployed into combat. There were, however, also critics of the vehicle, especially among high-ranking Romanian officials. It never saw action because the invading Soviet army had stopped its production. Other Romanian tank destroyers include the TACAM R-2 and TACAM T-60, which were converted from R-2 and T-60 light tanks respectively. Both of them saw action. One TACAM R-2 survives today and is displayed at the National Military Museum in Bucharest. Another conversion was the VDC R-35, Romania's only turreted tank destroyer. Two other proposed tank destroyers existed: the TACAM R-1 and TACAM T-38. Poland Variants of the Polish TKS and TK-3 tankettes up-armed with 20 mm gun (23–26 vehicles) were operationally deployed in the invasion of Poland. They were used as an anti-tank component of the reconnaissance units. There were also 37 mm armed TKS-D (2 experimental vehicles) and 47 mm armed TKD (4 experimental vehicles). It is not certain whether they were used operationally at all. France Due to the quick defeat of France, few French vehicles were built. The Laffly W15 TCC (Chasseur de chars) was an attempt to quickly build a light tank destroyer by mounting a 47 mm SA37 anti-tank gun onto a lightly armoured Laffly W15T artillery tractor. Other French tank destroyers were being developed, including the SOMUA SAu-40, ARL V39 and various ad hoc conversions of the Lorraine 37L. Subsequent developments In the face of the Warsaw Pact, a general need for extra firepower was identified. In the late 1960s, West Germany developed the Kanonenjagdpanzer, essentially a modernized World War II Jagdpanzer mounting a gun. As Soviet designs became more heavily armoured, the gun became ineffective and the Kanonenjagdpanzers were retrofitted for different roles or retired. Some provisions were made for the fitting of a 105 mm cannon, and many of the vehicles were modified to fire HOT or TOW missiles in place of a main gun. These upgraded variants remained in service into the 1990s. With the development of flexible anti-tank missiles, which were capable of installation on almost any vehicle in the 1960s, the concept of the tank destroyer has morphed into light vehicles with missiles. With the weight of main battle tanks growing to the forty to seventy-tonne range, airborne forces were unable to deploy reasonable anti-tank forces. The result was a number of attempts to make a light vehicle, including the conventional ASU-85, M56 Scorpion, the recoilless rifle-armed Ontos, and missile-armed Humber Hornet armoured truck and Sheridan light assault vehicle. The recent entries into that category are the 2S25 Sprut-SD, armed with a current-issue 125 mm tank gun that is also capable of launching missiles like the 9M119 Svir, and Israeli-modified Pandur IIs, which is to enter service with the Philippine Army by 2022 armed with an Elbit Turret and a 105 mm gun. Many forces' infantry fighting vehicles (IFVs) carry anti-tank missiles in every infantry platoon, and attack helicopters have also added anti-tank capability to the modern battlefield. But there are still dedicated anti-tank vehicles with very heavy long-range missiles, and ones intended for airborne use. There have also been dedicated anti-tank vehicles built on ordinary armoured personnel carrier or armored car chassis. Examples include the U.S. M901 ITV (Improved TOW Vehicle) and the Norwegian NM142, both on an M113 chassis, several Soviet ATGM launchers based on the BRDM reconnaissance car, the British FV438 Swingfire and FV102 Striker and the German Raketenjagdpanzer series built on the chassis of the HS 30 and Marder IFV. India fielded NAMIS (Nag Missile System) equipped with Nag Missiles. A US Army combined arms battalion has two infantry companies with TOW missile-armed Bradley IFVs and can bring a large concentration of accurate and lethal fire to bear on an attacking enemy unit that uses AFVs. They can be complemented by mobile units of AH-64 Apache helicopters armed with Hellfire antitank missiles. Missile carrying vehicles however are referred to as anti-tank missile carriers instead of tank destroyers. Some gun-armed tank destroyers remain in use. China has developed the tracked PTZ89 and the wheeled PTL02 tank destroyers. The PTZ89 is armed with a smoothbore cannon while the PTL02, developed by NORINCO for the PLA's new light (rapid reaction) mechanized infantry divisions, carries a one (a version armed with a 105 mm rifled gun is available for export). The PTL02 is built on the 6×6 wheeled chassis of the WZ551 APC. Italy and Spain use the Italian-built Centauro, a wheeled tank destroyer with a cannon. Russia, meanwhile, uses the Russian-built 2S25 Sprut-SD, operating as an amphibious light tank/tank destroyer armed with a 125mm cannon. The Sabrah Pandur II is a Wheeled tank destroyer variant of the Sabrah Light Tank developed by the Elbit Systems of Israel for the Philippine Army's future combat systems. See also Armoured warfare Self-propelled anti-aircraft weapon Self-propelled artillery Notes References Harry Yeide, (2005) The Tank Killers: A History of America's World War II Tank Destroyer Force. Havertown, PA: Casemate. . External links Tankdestroyer.net Popular Science, April 1940, Tanks Can Be Destroyed article on early US Army concepts for tank destroyers Tank Destroyer List ms:Pemusnah kereta kebal
2444
https://en.wikipedia.org/wiki/Conservation%20and%20restoration%20of%20cultural%20property
Conservation and restoration of cultural property
The conservation and restoration of cultural property focuses on protection and care of cultural property (tangible cultural heritage), including artworks, architecture, archaeology, and museum collections. Conservation activities include preventive conservation, examination, documentation, research, treatment, and education. This field is closely allied with conservation science, curators and registrars. Definition Conservation of cultural property involves protection and restoration using "any methods that prove effective in keeping that property in as close to its original condition as possible for as long as possible." Conservation of cultural heritage is often associated with art collections and museums and involves collection care and management through tracking, examination, documentation, exhibition, storage, preventive conservation, and restoration. The scope has widened from art conservation, involving protection and care of artwork and architecture, to conservation of cultural heritage, also including protection and care of a broad set of other cultural and historical works. Conservation of cultural heritage can be described as a type of ethical stewardship. It may broadly be divided into: Conservation and restoration of movable cultural property Conservation and restoration of immovable cultural property Conservation of cultural property applies simple ethical guidelines: Minimal intervention; Appropriate materials and reversible methods; Full documentation of all work undertaken. Often there are compromises between preserving appearance, maintaining original design and material properties, and ability to reverse changes. Reversibility is now emphasized so as to reduce problems with future treatment, investigation, and use. In order for conservators to decide upon an appropriate conservation strategy and apply their professional expertise accordingly, they must take into account views of the stakeholder, the values, artist's intent, meaning of the work, and the physical needs of the material. Cesare Brandi in his Theory of Restoration, describes restoration as "the methodological moment in which the work of art is appreciated in its material form and in its historical and aesthetic duality, with a view to transmitting it to the future". History and science Key dates Some consider the tradition of conservation of cultural heritage in Europe to have begun in 1565 with the restoration of the Sistine Chapel frescoes, but more ancient examples include the work of Cassiodorus. Brief history The care of cultural heritage has a long history, one that was primarily aimed at fixing and mending objects for their continued use and aesthetic enjoyment. Until the early 20th century, artists were normally the ones called upon to repair damaged artworks. During the 19th century, however, the fields of science and art became increasingly intertwined as scientists such as Michael Faraday began to study the damaging effects of the environment to works of art. Louis Pasteur carried out scientific analysis on paint as well. However, perhaps the first organized attempt to apply a theoretical framework to the conservation of cultural heritage came with the founding in the United Kingdom of the Society for the Protection of Ancient Buildings in 1877. The society was founded by William Morris and Philip Webb, both of whom were deeply influenced by the writings of John Ruskin. During the same period, a French movement with similar aims was being developed under the direction of Eugène Viollet-le-Duc, an architect and theorist, famous for his restorations of medieval buildings. Conservation of cultural heritage as a distinct field of study initially developed in Germany, where in 1888 Friedrich Rathgen became the first chemist to be employed by a Museum, the Koniglichen Museen, Berlin (Royal Museums of Berlin). He not only developed a scientific approach to the care of objects in the collections, but disseminated this approach by publishing a Handbook of Conservation in 1898. The early development of conservation of cultural heritage in any area of the world is usually linked to the creation of positions for chemists within museums. In British archaeology, key research and technical experimentation in conservation was undertaken by women such as Ione Gedye both in the field and in archaeological collections, particularly those of the Institute of Archaeology, London. In the United Kingdom, pioneering research into painting materials and conservation, ceramics, and stone conservation was conducted by Arthur Pillans Laurie, academic chemist and Principal of Heriot-Watt University from 1900. Laurie's interests were fostered by William Holman Hunt. In 1924 the chemist Harold Plenderleith began to work at the British Museum with Alexander Scott in the recently created Research Laboratory, although he was actually employed by the Department of Scientific and Industrial Research in the early years. Plenderleith's appointment may be said to have given birth to the conservation profession in the UK, although there had been craftsmen in many museums and in the commercial art world for generations. This department was created by the museum to address the deteriorating condition of objects in the collection, damages which were a result of their being stored in the London Underground tunnels during the First World War. The creation of this department moved the focus for the development of conservation theory and practice from Germany to Britain, and made the latter a prime force in this fledgling field. In 1956 Plenderleith wrote a significant handbook called The Conservation of Antiquities and Works of Art, which supplanted Rathgen's earlier tome and set new standards for the development of art and conservation science. In the United States, the development of conservation of cultural heritage can be traced to the Fogg Art Museum, and Edward Waldo Forbes, its director from 1909 to 1944. He encouraged technical investigation, and was Chairman of the Advisory Committee for the first technical journal, Technical Studies in the Field of the Fine Arts, published by the Fogg from 1932 to 1942. Importantly he also brought onto the museum staff chemists. Rutherford John Gettens was the first of such in the US to be permanently employed by an art museum. He worked with George L. Stout, the founder and first editor of Technical Studies. Gettens and Stout co-authored Painting Materials: A Short Encyclopaedia in 1942, reprinted in 1966. This compendium is still cited regularly. Only a few dates and descriptions in Gettens' and Stout's book are now outdated. George T. Oliver, of Oliver Brothers Art Restoration and Art Conservation-Boston (Est. 1850 in New York City) invented the vacuum hot table for relining paintings in 1920s; he filed a patent for the table in 1937. Taylor's prototype table, which he designed and constructed, is still in operation. Oliver Brothers is believed to be the first and the oldest continuously operating art restoration company in the United States. The focus of conservation development then accelerated in Britain and America, and it was in Britain that the first International Conservation Organisations developed. The International Institute for Conservation of Historic and Artistic Works (IIC) was incorporated under British law in 1950 as "a permanent organization to co-ordinate and improve the knowledge, methods, and working standards needed to protect and preserve precious materials of all kinds." The rapid growth of conservation professional organizations, publications, journals, newsletters, both internationally and in localities, has spearheaded the development of the conservation profession, both practically and theoretically. Art historians and theorists such as Cesare Brandi have also played a significant role in developing conservation science theory. In recent years ethical concerns have been at the forefront of developments in conservation. Most significantly has been the idea of preventive conservation. This concept is based in part on the pioneering work by Garry Thomson CBE, and his book Museum Environment, first published in 1978. Thomson was associated with the National Gallery in London; it was here that he established a set of guidelines or environmental controls for the best conditions in which objects could be stored and displayed within the museum environment. Although his exact guidelines are no longer rigidly followed, they did inspire this field of conservation. Conservation laboratories Conservators routinely use chemical and scientific analysis for the examination and treatment of cultural works. The modern conservation laboratory uses equipment such as microscopes, spectrometers, and various x-ray regime instruments to better understand objects and their components. The data thus collected helps in deciding the conservation treatments to be provided to the object. Ethics The conservator's work is guided by ethical standards. These take the form of applied ethics. Ethical standards have been established across the world, and national and international ethical guidelines have been written. One such example is: American Institute for Conservation Code of Ethics and Guidelines for Practice Conservation OnLine provides resources on ethical issues in conservation, including examples of codes of ethics and guidelines for professional conduct in conservation and allied fields; and charters and treaties pertaining to ethical issues involving the preservation of cultural property. As well as standards of practice conservators deal with wider ethical concerns, such as the debates as to whether all art is worth preserving. Keeping up with the international contemporary scenario, recent concerns with sustainability in conservation have emerged. The common understanding that "the care of an artifact should not come at the undue expense of the environment" is generally well accepted within the community and is already contemplated in guidelines of diverse institutions related to the field. Practice Preventive conservation Many cultural works are sensitive to environmental conditions such as temperature, humidity and exposure to visible light and ultraviolet radiation. These works must be protected in controlled environments where such variables are maintained within a range of damage-limiting levels. For example, watercolour paintings usually require shielding from sunlight to prevent fading of pigments. Collections care is an important element of museum policy. It is an essential responsibility of members of the museum profession to create and maintain a protective environment for the collections in their care, whether in store, on display, or in transit. A museum should carefully monitor the condition of collections to determine when an artifact requires conservation work and the services of a qualified conservator. Interventive conservation and restoration A teaching programme of interventive conservation was established in the UK at the Institute of Archaeology by Ione Gedye, which is still teaching interventive conservators today. A principal aim of a cultural conservator is to reduce the rate of deterioration of an object. Both non-interventive and interventive methodologies may be employed in pursuit of this goal. Interventive conservation refers to any direct interaction between the conservator and the material fabric of the object. Interventive actions are carried out for a variety of reasons, including aesthetic choices, stabilization needs for structural integrity, or cultural requirements for intangible continuity. Examples of interventive treatments include the removal of discolored varnish from a painting, the application of wax to a sculpture, and the washing and rebinding of a book. Ethical standards within the field require that the conservator fully justify interventive actions and carry out documentation before, during, and after the treatment. One of the guiding principles of conservation of cultural heritage has traditionally been the idea of reversibility, that all interventions with the object should be fully reversible and that the object should be able to be returned to the state in which it was prior to the conservator's intervention. Although this concept remains a guiding principle of the profession, it has been widely critiqued within the conservation profession and is now considered by many to be "a fuzzy concept." Another important principle of conservation is that all alterations should be well documented and should be clearly distinguishable from the original object. An example of a highly publicized interventive conservation effort would be the conservation work conducted on the Sistine Chapel. Example of an archaeological discovery and restoration of a mural painting Example of the restoration of an oil painting Sustainable conservation Recognising that conservation practices should not harm the environment, harm people, or contribute to global warming, the conservation-restoration profession has more recently focused on practices that reduce waste, reduce energy costs, and minimise the use of toxic or harmful solvents. A number of research projects, working groups, and other initiatives have explored how conservation can become a more environmentally sustainable profession. Sustainable conservation practices apply both to work within cultural institutions (e.g. museums, art galleries, archives, libraries, research centres and historic sites) as well as to businesses and private studios. Choice of materials Conservators and restorers use a wide variety of materials - in conservation treatments, and those used to safely transport, display and store cultural heritage items. These materials can include solvents, papers and boards, fabrics, adhesives and consolidants, plastics and foams, wood products, and many others. Stability and longevity are two important factors conservators consider when selecting materials; sustainability is becoming an increasingly important third. Examples of sustainable material choices and practices include: Using biodegradable products or those with less environmental impact where possible; Using 'green solvents' instead of more toxic alternatives, or treatment strategies that use much smaller amounts of solvents - for example, semi-rigid aqueous gels, emulsions or nano materials; Preparing smaller amounts of material (e.g. adhesives) to avoid waste; Observing recommended disposal protocols for chemicals, recyclable materials and compostable materials, particularly to avoid contamination of waterways; Choosing protective work wear that can be washed or cleaned and reused, rather than disposable options; Tracking stock quantities to avoid over-buying, especially for materials with expiration dates; Using durable materials for packing that may be washed and re-used, such as Tyvek or Mylar; Repurposing consumables such as blotting paper, non-woven fabrics, and polyester film when they are no longer fit for their original purpose; Using locally produced products whenever possible, to reduce carbon footprints; Reusing packaging materials such as cardboard boxes, plastic wrap and wooden crates; Using standard sizes of packaging and package designs that reduce waste; These decisions are not always straightforward - for example, installing deionised or distilled water filters in laboratories reduces waste associated with purchasing bottled products, but increases energy consumption. Similarly, locally-made papers and boards may reduce inherent carbon miles but they may be made with pulp sourced from old growth forests. Another dilemma is that many conservation-grade materials are chosen because they do not biodegrade. For example, when selecting a plastic with which to make storage enclosures, conservators prefer to use relatively long-lived plastics because they have better ageing properties - they are less likely to become yellow, leach plasticisers, or lose structural integrity and crumble (examples include polyethylene, polypropylene, and polyester). These plastics will also take longer to degrade in landfill. Energy use Many conservators and cultural organisations have sought to reduce the energy costs associated with controlling indoor storage and display environments (temperature, relative humidity, air filtration, and lighting levels) as well as those associated with the transport of cultural heritage items for exhibitions and loans. In general, lowering the temperature reduces the rate at which damaging chemical reactions occur within materials. For example, storing cellulose acetate film at 10 °C instead of 21 °C is estimated to increase its usable life by over 100 years. Controlling the relative humidity of air helps to reduce hydrolysis reactions and minimises cracking, distortion and other physical changes in hygroscopic materials. Changes in temperature will also bring about changes in relative humidity. Therefore, the conservation profession has placed great importance on controlling indoor environments. Temperature and humidity can be controlled through passive means (e.g. insulation, building design) or active means (air conditioning). Active controls typically require much higher energy use. Energy use increases with specificity - e.g. in will require more energy to maintain a quantity of air to a narrow temperature range (20-22 °C) than to a broad range (18-25 °C). In the past, conservation recommendations have often called for very tight, inflexible temperature and relative humidity set points. In other cases, conservators have recommended strict environmental conditions for buildings that could not reasonably be expected to achieve them, due to the quality of build, local environmental conditions (e.g. recommending temperate conditions for a building located in the tropics) or the financial circumstances of the organisation. This has been an area of particular debate for cultural heritage organisations who lend and borrow cultural items to each other - often, the lender will specify strict environmental conditions as part of the loan agreement, which may be very expensive for the borrowing organisation to achieve, or impossible. The energy costs associated with cold storage and digital storage are also gaining more attention. Cold storage is a very effective strategy to preserve at-risk collections such as cellulose nitrate and cellulose acetate film, which can deteriorate beyond use within decades at ambient conditions. Digital storage costs are rising for both born-digital cultural heritage (photographs, audiovisual, time-based media) and to store digital preservation and access copies of cultural heritage. Digital storage capacity is a major factor in the complexity of preserving digital heritage such as video games, social media, messaging services, and email. Other areas where energy use can be reduced within conservation and restoration include: Exhibition lighting - e.g. using lower-energy LED lighting systems and light sensors that switch lights on only when visitors are present; Installation of green energy capture systems in cultural organisations, such as solar photovoltaic plates, wind energy systems, and heat pumps; Improving the energy performance of cultural buildings by installing insulation, sealing gaps, reducing the number of windows and installing double-glazing: Using microclimates to house small groups of climate-sensitive objects instead of seeking to control the environmental conditions of the whole building. Country by country look United States Heritage Preservation, in partnership with the Institute of Museum and Library Services, a U.S. federal agency, produced The Heritage Health Index. The results of this work was the report A Public Trust at Risk: The Heritage Health Index Report on the State of America's Collections, which was published in December 2005 and concluded that immediate action is needed to prevent the loss of 190 million artifacts that are in need of conservation treatment. The report made four recommendations: Institutions must give priority to providing safe conditions for the collections they hold in trust. Every collecting institution must develop an emergency plan to protect its collections and train staff to carry it out. Every institution must assign responsibility for caring for collections to members of its staff. Individuals at all levels of government and in the private sector must assume responsibility for providing the support that will allow these collections to survive. United Kingdom In October 2006, the Department for Culture, Media and Sport, a governmental department, authored a document: "Understanding the Future: Priorities for England's Museums". This document was based on several years of consultation aimed to lay out the government's priorities for museums in the 21st century. The document listed the following as priorities for the next decade: Museums will fulfil their potential as learning resources (pp 7–10). Museums will be embedded into the delivery of education in every school in the country. Understanding of the effectiveness of museum education will be improved further and best practice built into education programmes. The value of museums' collections as a research resource will be well understood and better links built between the academic community and museums. Museums will embrace their role in fostering, exploring, celebrating and questioning the identities of diverse communities (pp 11–14). The sector needs to work with partners in academia and beyond to create an intellectual framework supporting museums' capacity to tackle issues of identity. The museum sector must continue to develop improved practical techniques for engaging communities of all sorts. Museums' collections will be more dynamic and better used (pp 15–18). Government and the sector will find new ways to encourage museums to collect actively and strategically, especially the record of contemporary society. The sector will develop new collaborative approaches to sharing and developing collections and related expertise. Museums' workforce will be dynamic, highly skilled and representative (pp 17–22). Museums' governing bodies and workforce will be representative of the communities they serve. Find more varied ways for a broader range of skills to come into museums. Improve continuing professional development. Museums will work more closely with each other and partners outside the sector (pp 23–26). A consistent evidence base of the contribution of all kinds of museums to the full range of public service agendas will be developed. There will be deeper and longer lasting partnerships between the national museums and a broader range of regional partners. Museums' international roles will be strengthened to improve museum programmes in this country and Britain's image, reputation and relationships abroad. The conservation profession response to this report was on the whole less than favourable, the Institute of Conservation (ICON) published their response under the title "A Failure of Vision". It had the following to say: Concluding: Further to this the ICON website summary report lists the following specific recommendations: A national survey to find out what the public want from museums, what motivates them to visit them and what makes for a rewarding visit. A review of survey results and prioritisation of the various intrinsic, instrumental and institutional values to provide a clear basis for a 10-year strategy HR consultants to be brought in from the commercial sector to review recruitment, career development and working practices in the national and regional museums. A commitment to examine the potential for using Museum Accreditation as a more effective driver for improving recruitment, diversity, and career development across the sector. DCMS to take full account of the eventual findings of the current Commons Select Committee enquiry into Care of Collections in the final version of this document The adoption of those recommendations of the recent House of Lords inquiry into Science and Heritage which might affect the future of museums. In November 2008, the UK-based think tank Demos published an influential pamphlet entitled It's a material world: caring for the public realm, in which they argue for integrating the public directly into efforts to conserve material culture, particularly that which is in the public, their argument, as stated on page 16, demonstrates their belief that society can benefit from conservation as a paradigm as well as a profession: Training Training in conservation of cultural heritage for many years took the form of an apprenticeship, whereby an apprentice slowly developed the necessary skills to undertake their job. For some specializations within conservation this is still the case. However, it is more common in the field of conservation today that the training required to become a practicing conservator comes from a recognized university course in conservation of cultural heritage. The university can rarely provide all the necessary training in first hand experience that an apprenticeship can, and therefore in addition to graduate level training the profession also tends towards encouraging conservation students to spend time as an intern. Conservation of cultural heritage is an interdisciplinary field as conservators have backgrounds in the fine arts, sciences (including chemistry, biology, and materials science), and closely related disciplines, such as art history, archaeology, and anthropology. They also have design, fabrication, artistic, and other special skills necessary for the practical application of that knowledge. Within the various schools that teach conservation of cultural heritage, the approach differs according to the educational and vocational system within the country, and the focus of the school itself. This is acknowledged by the American Institute for Conservation who advise "Specific admission requirements differ and potential candidates are encouraged to contact the programs directly for details on prerequisites, application procedures, and program curriculum". In France, training for heritage conservation is taught by four schools : , L'École supérieure des Beaux-Arts Tours, Angers, Le Mans, L'Université Paris 1 Panthéon-Sorbonne, Institut national du patrimoine. Associations and professional organizations Societies devoted to the care of cultural heritage have been in existence around the world for many years. One early example is the founding in 1877 of the Society for the Protection of Ancient Buildings in Britain to protect the built heritage, this society continues to be active today. The 14th Dalai Lama and the Tibetan people work to preserve their cultural heritage with organizations including the Tibetan Institute of Performing Arts and an international network of eight Tibet Houses. The built heritage was at the forefront of the growth of member based organizations in the United States. Preservation Virginia, founded in Richmond in 1889 as the Association for the Preservation of Virginia Antiquities, was the United States' first statewide historic preservation group. Today, professional conservators join and take part in the activities of numerous conservation associations and professional organizations with the wider field, and within their area of specialization. In Europe, E.C.C.O. European Confederation of Conservator-Restorers Organisations was established in 1991 by 14 European Conservator-Restorers' Organisations. Currently representing close to 6.000 professionals within 23 countries and 26 members organisations, including one international body (IADA), E.C.C.O. embodies the field of preservation of cultural heritage, both movable and immovable. These organizations exist to "support the conservation professionals who preserve our cultural heritage". This involves upholding professional standards, promoting research and publications, providing educational opportunities, and fostering the exchange of knowledge among cultural conservators, allied professionals, and the public. International cultural property documents See also Conservation and restoration of rail vehicles The Georgian Group Wikipedia:WikiProject Collections Care International Day For Monuments and Sites References Further reading Copies of this volume are available for free pdf download from the Smithsonian's digital library by clicking on the included link. External links BCIN, the Bibliographic Database of the Conservation Information Network CAMEO: Conservation and Art Materials Encyclopedia OnLine Conservation OnLine (CoOL) Resources for Conservation Professionals DOCAM — Documentation and Conservation of the Media Arts Heritage ICOMOS Open Archive: EPrints on Cultural Heritage Publications & Resources at the Getty Conservation Institute Art history Museology Cultural heritage Articles containing video clips Cultural heritage conservation
2627
https://en.wikipedia.org/wiki/Aegina
Aegina
Aegina (; , Aígina ; ) is one of the Saronic Islands of Greece in the Saronic Gulf, from Athens. Tradition derives the name from Aegina, the mother of the hero Aeacus, who was born on the island and became its king. Administration Municipality The municipality of Aegina consists of the island of Aegina and a few offshore islets. It is part of the Islands regional unit, Attica region. The municipality is subdivided into the following five communities (population in 2011 in parentheses ): Kypseli (2,124) Mesagros (1,361) Perdika (8,236) Vathy (1,495) The regional capital is the town of Aegina, situated at the northwestern end of the island. Due to its proximity to Athens, it is a popular vacation place during the summer months, with quite a few Athenians owning second houses on the island. The buildings of the island are examples of Neoclassical architecture with a strong folk element, built in the 19th century Province The province of Aegina () was one of the provinces of the Attica Prefecture and was created in 1833 as part of Attica and Boeotia Prefecture. Its territory corresponded with that of the current municipalities Aegina and Agkistri. It was abolished in 2006. Geography Aegina is roughly triangular in shape, approximately from east to west and from north to south, with an area of . An extinct volcano constitutes two-thirds of Aegina. The northern and western sides consist of stony but fertile plains, which are well cultivated and produce luxuriant crops of grain, with some cotton, vines, almonds, olives and figs, but the most characteristic crop of Aegina today (2000s) is pistachio. Economically, the sponge fisheries are of notable importance. The southern volcanic part of the island is rugged and mountainous, and largely barren. Its highest rise is the conical Mount Oros (531 m) in the south, and the Panhellenian ridge stretches northward with narrow fertile valleys on either side. The beaches are also a popular tourist attraction. Hydrofoil ferries from Piraeus take only forty minutes to reach Aegina; the regular ferry takes about an hour, with ticket prices for adults within the 4–15 euro range. There are regular bus services from Aegina town to destinations throughout the island such as Agia Marina. Portes is a fishing village on the east coast. Climate Aegina has a hot semi-arid climate (Köppen climate classification: BSh). It is one of the driest places in Greece. History Earliest history (20th–7th centuries BC) Aegina, according to Herodotus, was a colony of Epidaurus, to which state it was originally subject. Its placement between Attica and the Peloponnesus made it a site of trade even earlier, and its earliest inhabitants allegedly came from Asia Minor. Minoan ceramics have been found in contexts of . The famous Aegina Treasure, now in the British Museum is estimated to date between 1700 and 1500 BC. The discovery on the island of a number of gold ornaments belonging to the last period of Mycenaean art suggests that Mycenaean culture existed in Aegina for some generations after the Dorian conquest of Argos and Lacedaemon. It is probable that the island was not Doricised before the 9th century BC. One of the earliest historical facts is its membership in the Amphictyony or League of Calauria, attested around the 8th century BC. This ostensibly religious league included, besides Aegina, Athens, the Minyan (Boeotian) Orchomenos, Troezen, Hermione, Nauplia, and Prasiae. It was probably an organisation of city-states that were still Mycenaean, for the purpose of suppressing piracy in the Aegean that began as a result of the decay of the naval supremacy of the Mycenaean princes. Aegina seems to have belonged to the Eretrian league during the Lelantine War; this, perhaps, may explain the war with Samos, a major member of the rival Chalcidian league during the reign of King Amphicrates (Herod. iii. 59), i.e. not later than the earlier half of the 7th century BC. Coinage and sea power (7th–5th centuries BC) Its early history reveals that the maritime importance of the island dates back to pre-Dorian times. It is usually stated on the authority of Ephorus, that Pheidon of Argos established a mint in Aegina, the first city-state to issue coins in Europe, the Aeginetic stater. One stamped stater (having the mark of some authority in the form of a picture or words) can be seen in the Bibliothèque Nationale of Paris. It is an electrum stater of a turtle, an animal sacred to Aphrodite, struck at Aegina that dates from 700 BC. Therefore, it is thought that the Aeginetes, within 30 or 40 years of the invention of coinage in Asia Minor by the Ionian Greeks or the Lydians (c. 630 BC), might have been the ones to introduce coinage to the Western world. The fact that the Aeginetic standard of weights and measures (developed during the mid-7th century) was one of the two standards in general use in the Greek world (the other being the Euboic-Attic) is sufficient evidence of the early commercial importance of the island. The Aeginetic weight standard of about 12.2 grams was widely adopted in the Greek world during the 7th century BC. The Aeginetic stater was divided into two drachmae of 6.1 grams of silver. Staters depicting a sea-turtle were struck up to the end of the 5th century BC. During the First Peloponnesian War, by 456 BC, it was replaced by the land tortoise. During the naval expansion of Aegina during the Archaic Period, Kydonia was an ideal maritime stop for Aegina's fleet on its way to other Mediterranean ports controlled by the emerging sea-power Aegina. During the next century Aegina was one of the three principal states trading at the emporium of Naucratis in Egypt, and it was the only Greek state near Europe that had a share in this factory. At the beginning of the 5th century BC it seems to have been an entrepôt of the Pontic grain trade, which, at a later date, became an Athenian monopoly. Unlike the other commercial states of the 7th and 6th centuries BC, such as Corinth, Chalcis, Eretria and Miletus, Aegina did not found any colonies. The settlements to which Strabo refers (viii. 376) cannot be regarded as any real exceptions to this statement. Rivalry with Athens (5th century BC) The known history of Aegina is almost exclusively a history of its relations with the neighbouring state of Athens, which began to compete with the thalassocracy (sea power) of Aegina about the beginning of the 6th century BC. Solon passed laws limiting Aeginetan commerce in Attica. The legendary history of these relations, as recorded by Herodotus (v. 79–89; vi. 49–51, 73, 85–94), involves critical problems of some difficulty and interest. He traces the hostility of the two states back to a dispute about the images of the goddesses Damia and Auxesia, which the Aeginetes had carried off from Epidauros, their parent state. The Epidaurians had been accustomed to make annual offerings to the Athenian deities Athena and Erechtheus in payment for the Athenian olive-wood of which the statues were made. Upon the refusal of the Aeginetes to continue these offerings, the Athenians endeavoured to carry away the images. Their design was frustrated miraculously (according to the Aeginetan version, the statues fell upon their knees) and only a single survivor returned to Athens. There he became victim to the fury of his comrades' widows who pierced him with their peplos brooch-pins. No date is assigned by Herodotus for this "old feud"; recent writers, such as J. B. Bury and R. W. Macan, suggest the period between Solon and Peisistratus, . It is possible that the whole episode is mythical. A critical analysis of the narrative seems to reveal little else than a series of aetiological traditions (explanatory of cults and customs), such as of the kneeling posture of the images of Damia and Auxesia, of the use of native ware instead of Athenian in their worship, and of the change in women's dress at Athens from the Dorian peplos to the Ionian style chiton. In the early years of the 5th century BC the Thebans, after the defeat by Athens about 507 BC, appealed to Aegina for assistance. The Aeginetans at first contented themselves with sending the images of the Aeacidae, the tutelary heroes of their island. Subsequently, however, they contracted an alliance, and ravaged the seaboard of Attica. The Athenians were preparing to make reprisals, in spite of the advice of the Delphic oracle that they should desist from attacking Aegina for thirty years, and content themselves meanwhile with dedicating a precinct to Aeacus, when their projects were interrupted by the Spartan intrigues for the restoration of Hippias. In 491 BC Aegina was one of the states which gave the symbols of submission ("earth and water") to Achaemenid Persia. Athens at once appealed to Sparta to punish this act of medism, and Cleomenes I, one of the Spartan kings, crossed over to the island, to arrest those who were responsible for it. His attempt was at first unsuccessful; but, after the deposition of Demaratus, he visited the island a second time, accompanied by his new colleague Leotychides, seized ten of the leading citizens and deposited them at Athens as hostages. After the death of Cleomenes and the refusal of the Athenians to restore the hostages to Leotychides, the Aeginetes retaliated by seizing a number of Athenians at a festival at Sunium. Thereupon the Athenians concerted a plot with Nicodromus, the leader of the democratic party in the island, for the betrayal of Aegina. He was to seize the old city, and they were to come to his aid on the same day with seventy vessels. The plot failed owing to the late arrival of the Athenian force, when Nicodromus had already fled the island. An engagement followed in which the Aeginetes were defeated. Subsequently, however, they succeeded in winning a victory over the Athenian fleet. All the incidents subsequent to the appeal of Athens to Sparta are referred expressly by Herodotus to the interval between the sending of the heralds in 491 BC and the invasion of Datis and Artaphernes in 490 BC (cf. Herod. vi. 49 with 94). There are difficulties with this story, of which the following are the principal elements: Herodotus nowhere states or implies that peace was concluded between the two states before 481 BC, nor does he distinguish between different wars during this period. Hence it would follow that the war lasted from soon after 507 BC until the congress at the Isthmus of Corinth in 481 BC It is only for two years (491 and 490 BC) out of the twenty-five that any details are given. It is the more remarkable that no incidents are recorded in the period between the battles of Marathon and Salamis, since at the time of the Isthmian Congress the war was described as the most important one then being waged in Greece, It is improbable that Athens would have sent twenty vessels to the aid of the Ionians in 499 BC if at the time it was at war with Aegina. There is an incidental indication of time, which indicates the period after Marathon as the true date for the events which are referred by Herodotus to the year before Marathon, viz. the thirty years that were to elapse between the dedication of the precinct to Aeacus and the final victory of Athens. As the final victory of Athens over Aegina was in 458 BC, the thirty years of the oracle would carry us back to the year 488 BC as the date of the dedication of the precinct and the beginning of hostilities. This inference is supported by the date of the building of the 200 triremes "for the war against Aegina" on the advice of Themistocles, which is given in the Constitution of Athens as 483–482 BC. It is probable, therefore, that Herodotus is in error both in tracing back the beginning of hostilities to an alliance between Thebes and Aegina () and in claiming the episode of Nicodromus occurred prior to the battle of Marathon. Overtures were unquestionably made by Thebes for an alliance with Aegina , but they came to nothing. The refusal of Aegina was in the diplomatic guise of "sending the Aeacidae." The real occasion of the beginning of the war was the refusal of Athens to restore the hostages some twenty years later. There was but one war, and it lasted from 488 to 481 BC. That Athens had the worst of it in this war is certain. Herodotus had no Athenian victories to record after the initial success, and the fact that Themistocles was able to carry his proposal to devote the surplus funds of the state to the building of so large a fleet seems to imply that the Athenians were themselves convinced that a supreme effort was necessary. It may be noted, in confirmation of this opinion, that the naval supremacy of Aegina is assigned by the ancient writers on chronology to precisely this period, i.e. the years 490–480 BC. Decline In the repulse of Xerxes I it is possible that the Aeginetes played a larger part than is conceded to them by Herodotus. The Athenian tradition, which he follows in the main, would naturally seek to obscure their services. It was to Aegina rather than Athens that the prize of valour at Salamis was awarded, and the destruction of the Persian fleet appears to have been as much the work of the Aeginetan contingent as of the Athenian (Herod. viii. 91). There are other indications, too, of the importance of the Aeginetan fleet in the Greek scheme of defence. In view of these considerations it becomes difficult to credit the number of the vessels that is assigned to them by Herodotus (30 as against 180 Athenian vessels, cf. Greek History, sect. Authorities). During the next twenty years the Philo-Laconian policy of Cimon secured Aegina, as a member of the Spartan league, from attack. The change in Athenian foreign policy, which was consequent upon the ostracism of Cimon in 461 BC, resulted in what is sometimes called the First Peloponnesian War, during which most of the fighting was experienced by Corinth and Aegina. The latter state was forced to surrender to Athens after a siege, and to accept the position of a subject-ally (). The tribute was fixed at 30 talents. By the terms of the Thirty Years' Peace (445 BC) Athens promised to restore to Aegina her autonomy, but the clause remained ineffective. During the first winter of the Peloponnesian War (431 BC) Athens expelled the Aeginetans and established a cleruchy in their island. The exiles were settled by Sparta in Thyreatis, on the frontiers of Laconia and Argolis. Even in their new home they were not safe from Athenian rancour. A force commanded by Nicias landed in 424 BC, and killed most of them. At the end of the Peloponnesian War Lysander restored the scattered remnants of the old inhabitants to the island, which was used by the Spartans as a base for operations against Athens during the Corinthian War. It is probable that the power of Aegina had steadily declined during the twenty years after Salamis, and that it had declined absolutely, as well as relatively to that of Athens. Commerce was the source of Aegina's greatness, and her trade, which seems to have been principally with the Levant, must have suffered seriously from the war with Persia. Aegina's medism in 491 is to be explained by its commercial relations with the Persian Empire. It was forced into patriotism in spite of itself, and the glory won by the Battle of Salamis was paid for by the loss of its trade and the decay of its marine. The loss of the state's power is explained by the conditions of the island, which was based on slave labour; Aristotle's estimated the population of slaves were as much as 470,000. Hellenistic period and Roman rule Aegina with the rest of Greece became dominated successively by the Macedonians (322–229 BC), the Achaeans (229–211 BC), Aetolians (211–210  BC), Attalus of Pergamum (210–133 BC) and the Romans (after 133 BC). A sign at the Archaeological Museum of Aegina is reported to say that a Jewish community was established in Aegina "at the end of the second and during the 3rd century AD" by Jews fleeing the barbarian invasions of the time in Greece. However, the first phases of those invasions began in the 4th century. The Romaniote jewish community erected an elaborate synagogue in rectangle form with an apse on the eastern wall with a magnificent mosaic decorated with geometric motifs, still preserved in the courtyard of the Archaeological Museum of Aegina. The synagogue dates from the 4th century AD and was in use until the 7th century AD. Local Christian tradition has it that a Christian community was established there in the 1st century, having as its bishop Crispus, the ruler of the Corinthian synagogue, who became a Christian, and was baptised by Paul the Apostle. There are written records of participation by later bishops of Aegina, Gabriel and Thomas, in the Councils of Constantinople in 869 and 879. The see was at first a suffragan of the metropolitan see of Corinth, but was later given the rank of archdiocese. No longer a residential bishopric, Aegina is today listed by the Catholic Church as a titular see. Byzantine period Aegina belonged to the East Roman (Byzantine) Empire after the division of the Roman Empire in 395. It remained Eastern Roman during the period of crisis of the 7th–8th centuries, when most of the Balkans and the Greek mainland were overrun by Slavic invasions. Indeed, according to the Chronicle of Monemvasia, the island served as a refuge for the Corinthians fleeing these incursions. The island flourished during the early 9th century, as evidenced by church construction activity, but suffered greatly from Arab raids originating from Crete. Various hagiographies record a large-scale raid , that resulted in the flight of much of the population to the Greek mainland. During that time, some of the population sought refuge in the island's hinterland, establishing the settlement of Palaia Chora. According to the 12th-century bishop of Athens, Michael Choniates, by his time the island had become a base for pirates. This is corroborated by Benedict of Peterborough's graphic account of Greece, as it was in 1191; he states that many of the islands were uninhabited for fear of pirates and that Aegina, along with Salamis and Makronisos, were their strongholds. Frankish rule after 1204 After the dissolution and partition of the Byzantine Empire by the Fourth Crusade in 1204, Aegina was accorded to the Republic of Venice. In the event, it became controlled by the Duchy of Athens. The Catalan Company seized control of Athens, and with it Aegina, in 1317, and in 1425 the island became controlled by the Venetians, when Alioto Caopena, at that time ruler of Aegina, placed himself by treaty under the Republic's protection to escape the danger of a Turkish raid. The island must then have been fruitful, for one of the conditions by which Venice accorded him protection was that he should supply grain to Venetian colonies. He agreed to surrender the island to Venice if his family became extinct. Antonio II Acciaioli opposed the treaty for one of his adopted daughters had married the future lord of Aegina, Antonello Caopena. Venetians in Aegina (1451–1537) In 1451, Aegina became Venetian. The islanders welcomed Venetian rule; the claims of Antonello's uncle Arnà, who had lands in Argolis, were satisfied by a pension. A Venetian governor (rettore) was appointed, who was dependent on the authorities of Nauplia. After Arnà's death, his son Alioto renewed his claim to the island but was told that the republic was resolved to keep it. He and his family were pensioned and one of them aided in the defence of Aegina against the Turks in 1537, was captured with his family, and died in a Turkish dungeon. In 1463 the Turco-Venetian war began, which was destined to cost the Venetians Negroponte (Euboea), the island of Lemnos, most of the Cyclades islands, Scudra and their colonies in the Morea. Peace was concluded in 1479. Venice still retained Aegina, Lepanto (Naupactus), Nauplia, Monemvasia, Modon, Navarino, Coron, and the islands Crete, Mykonos and Tinos. Aegina remained subject to Nauplia. Administration Aegina obtained money for its defences by reluctantly sacrificing its cherished relic, the head of St. George, which had been carried there from Livadia by the Catalans. In 1462, the Venetian Senate ordered the relic to be removed to St. Giorgio Maggiore in Venice and on 12 November, it was transported from Aegina by Vettore Cappello, the famous Venetian commander. In return, the Senate gave the Aeginetes 100 ducats apiece towards fortifying the island. In 1519, the government was reformed. The system of having two rectors was found to result in frequent quarrels and the republic thenceforth sent out a single official styled Bailie and Captain, assisted by two councillors, who performed the duties of camerlengo by turns. The Bailie's authority extended over the rector of Aegina, whereas Kastri (opposite the island Hydra) was granted to two families, the Palaiologoi and the Alberti. Society at Nauplia was divided into three classes: nobles, citizens and plebeians, and it was customary for nobles alone to possess the much-coveted local offices, such as the judge of the inferior court and inspector of weights and measures. The populace now demanded its share and the home government ordered that at least one of the three inspectors should be a non-noble. Aegina had always been exposed to the raids of corsairs and had oppressive governors during these last 30 years of Venetian rule. Venetian nobles were not willing to go to this island. In 1533, three rectors of Aegina were punished for their acts of injustice and there is a graphic account of the reception given by the Aeginetans to the captain of Nauplia, who came to command an enquiry into the administration of these delinquents (vid. inscription over the entrance of St. George the Catholic in Paliachora). The rectors had spurned their ancient right to elect an islander to keep one key of the money-chest. They had also threatened to leave the island en masse with the commissioner, unless the captain avenged their wrongs. To spare the economy of the community, it was ordered that appeals from the governor's decision should be made on Crete, instead of in Venice. The republic was to pay a bakshish to the Turkish governor of the Morea and to the voivode who was stationed at the frontier of Thermisi (opposite Hydra). The fortifications too, were allowed to become decrepit and were inadequately guarded. 16th century After the end of the Duchy of Athens and the principality of Achaia, the only Latin possessions left on the mainland of Greece were the papal city of Monemvasia, the fortress of Vonitsa, the Messenian stations Coron and Modon, Lepanto, Pteleon, Navarino, and the castles of Argos and Nauplia, to which the island of Aegina was subordinate. In 1502–03, the new peace treaty left Venice with nothing but Cephalonia, Monemvasia and Nauplia, with their appurtenances in the Morea. And against the sack of Megara, it had to endure the temporary capture of the castle of Aegina by Kemal Reis and the abduction of 2000 inhabitants. This treaty was renewed in 1513 and 1521. All supplies of grain from Nauplia and Monemvasia had to be imported from Turkish possessions, while corsairs rendered dangerous all traffic by sea. In 1537, sultan Suleiman declared war upon Venice and his admiral Hayreddin Barbarossa devastated much of the Ionian Islands, and in October invaded the island of Aegina. On the fourth day Palaiochora was captured, but the Latin church of St George was spared. Hayreddin Barbarossa had the adult male population massacred and took away 6,000 surviving women and children as slaves. Then Barbarossa sailed to Naxos, whence he carried off an immense booty, compelling the Duke of Naxos to purchase his further independence by paying a tribute of 5000 ducats. With the peace of 1540, Venice ceded Nauplia and Monemvasia. For nearly 150 years afterwards, Venice ruled no part of the mainland of Greece except Parga and Butrinto (subordinate politically to the Ionian Islands), but it still retained its insular dominions Cyprus, Crete, Tenos and six Ionian islands. First Ottoman period (1540–1687) The island was attacked and left desolate by Francesco Morosini during the Cretan War (1654). Second Venetian period (1687–1715) In 1684, the beginning of the Morean War between Venice and the Ottoman Empire resulted in the temporary reconquest of a large part of the country by the Republic. In 1687 the Venetian army arrived in Piraeus and captured Attica. The number of the Athenians at that time exceeded 6,000, the Albanians from the villages of Attica excluded, whilst in 1674 the population of Aegina did not seem to exceed 3,000 inhabitants, two thirds of which were women. The Aeginetans had been reduced to poverty to pay their taxes. The most significant plague epidemic began in Attica during 1688, an occasion that caused the massive migration of Athenians toward the south; most of them settled in Aegina. In 1693 Morosini resumed command, but his only acts were to refortify the castle of Aegina, which he had demolished during the Cretan war in 1655, the cost of upkeep being paid as long as the war lasted by the Athenians, and to place it and Salamis under Malipiero as Governor. This caused the Athenians to send him a request for the renewal of Venetian protection and an offer of an annual tribute. He died in 1694 and Zeno was appointed at his place. In 1699, thanks to English mediation, the war ended with the peace of Karlowitz by which Venice retained possession of the 7 Ionian islands as well as Butrinto and Parga, the Morea, Spinalonga and Suda, Tenos, Santa Maura and Aegina and ceased to pay a tribute for Zante, but which restored Lepanto to the Ottoman sultan. Cerigo and Aegina were united administratively since the peace with Morea, which not only paid all the expenses of administration but furnished a substantial balance for the naval defence of Venice, in which it was directly interested. Second Ottoman period (1715–1821) During the early part of the Ottoman–Venetian War of 1714–1718 the Ottoman Fleet commanded by Canum Hoca captured Aegina. Ottomans rule in Aegina and the Morea was resumed and confirmed by the Treaty of Passarowitz, and they retained control of the island with the exception of a brief Russian occupation Orlov Revolt (early 1770s), until the beginning of the Greek War of Independence in 1821. Greek Revolution During the Greek War of Independence, Aegina became an administrative centre for the Greek revolutionary authorities. Ioannis Kapodistrias was briefly established here. Landmarks Temple of Aphaea, dating from about 490 BC, it is the oldest surviving temple in Greece. It was dedicated to its namesake, a goddess who was later associated with Athena; the temple was part of an equilateral holy triangle of temples including the Athenian Parthenon and the temple of Poseidon at Sounion. Monastery of Agios Nectarios, dedicated to Nectarios of Aegina, a recent saint of the Greek Orthodox Church. A statue in the principal square commemorates Ioannis Kapodistrias (1776–1831), the first administrator of free modern Greece. The Orphanage of Kapodistrias is a large building, known locally as The Prison (Οι Φυλακές, Oi Filakes), constructed in 1828-29 by Ioannis Kapodistrias as a home for children orphaned as a result of the Greek War of Independence. The building also housed schools, vocational workshops, the National Public Library, the National Archaeological Museum, a military academy, the National Printing Office and the National Conservatory for Choir and Orchestra. From about 1880 it was used as a prison, and housed political prisoners during the Greek Junta (1967-1974) - hence its local name. There are currently plans to restore the building as a museum. The Tower of Markellos was probably built during the second Venetian occupation, 1687–1714, as a watch tower in anticipation of a Turkish siege. A castle, fortified walls and numerous watchtowers were built at this time. The tower was abandoned after the Turkish occupation of 1714, until revolutionary leader Spyros Markellos bought the tower as his residence in around 1802. In 1826-28 it was the headquarters of the temporary government of the embryonic Greek state. It subsequently was used as a police headquarters and housed various government agencies until it was abandoned again in the mid 19th century. It is currently owned by the Municipality of Aegina. Temple of Zeus Hellanios, near the village of Pachia Rachi, is a 13th-century Byzantine church, built on the ruins of the ancient temple to Zeus Hellanios, built in the 4th century BC. The staircase leading up to the church, some of the original walls, and loose stones from the earlier temple remain. Colona, Located to the north of the town of Aegina. Acropolis with the sanctuary of Apollo and Byzantine settlement. The name Colona was given by the Venetian sailors, who used the columns of the pavilion of the Doric temple of Apollo (6x11 columns) as a sign of orientation. The foundations and one column from the rear building are preserved. The temple with the buildings related to the function of the sanctuary dominates the ancient acropolis on the hill. It was built at the end of the 6th century when Aegina, one of the most important commercial centers, emerged as a rival of Athens. Excavations from the 19th century onwards made it clear that the architectural remains of the archaic-Hellenistic acropolis, which are only partially preserved, are based on the impressive buildings of the prehistoric era, with at least ten successive building phases. Economy Pistachios In 1896, the physician Nikolaos Peroglou introduced the systematic cultivation of pistachios, which soon became popular among the inhabitants of the island. By 1950, pistachio cultivation had significantly displaced the rest of the agricultural activity due to its high profitability but also due to the phylloxera that threatened the vineyards that time. As a result, in the early 60s, the first pistachio peeling factory was established in the Plakakia area by Grigorios Konidaris. The quality of "Fistiki Aeginis" (Aegina Pistachios), a name that was established as a product of Protected Designation of Origin (PDO) in 1996, is considered internationally excellent and superior to several foreign varieties, due to the special climatic conditions of the island (drought) as well as soil's volcanic characteristics. Pistachios have made Aegina famous all over the world. Today, half of the pistachio growers are members of the Agricultural Cooperative of Aegina's Pistachio Producers. It is estimated that pistachio cultivation covers 29,000 acres of the island while the total production reaches 2,700 tons per year. In recent years, in mid-September, the Pistachio Festival has been organized every year under the name "Fistiki Fest". Culture Mythology In Greek mythology, Aegina was a daughter of the river god Asopus and the nymph Metope. She bore at least two children: Menoetius by Actor, and Aeacus by the god Zeus. When Zeus abducted Aegina, he took her to Oenone, an island close to Attica. Here, Aegina gave birth to Aeacus, who would later become king of Oenone; thenceforth, the island's name was Aegina. Aegina was the gathering place of Myrmidons; in Aegina they gathered and trained. Zeus needed an elite army and at first thought that Aegina, which at the time did not have any villagers, was a good place. So he changed some ants (, Myrmigia) into warriors who had six hands and wore black armour. Later, the Myrmidons, commanded by Achilles, were known as the most fearsome fighting unit in Greece. Famous Aeginetans Aeacus, the first king of Aegina according to mythology, in whose honour the Aeacea were celebrated Smilis (6th century BC), sculptor Sostratus of Aegina (6th century BC), merchant Onatas (5th century BC), sculptor Ptolichus (5th century BC), sculptor Philiscus of Aegina (4th century BC), Cynic philosopher Paul of Aegina (7th century), medical scholar and physician Saint Athanasia of Aegina (9th century), abbess and saint Cosmas II Atticus (12th century), Patriarch of Constantinople Nectarios of Aegina (1846–1920), bishop and saint Aristeidis Moraitinis (aviator) born 1891, died 1918 Gustav Hasford, American military journalist and novelist, moved to Aegina and died there of heart failure on 29 January 1993, aged 45 Historical population See also Flag of Aegina Gallery References Sources Welter Gabriel, Aigina, Archäol. Inst. d. Deutschen Reiches, Berlin 1938. Miller William, Essays on the Latin orient, Rome 1921 (reprint: Amsterdam 1964).Essays on the Latin Orient Miller William, "Η Παληαχώρα της Αιγίνης. Ηρημωμένη ελληνική πόλις", Νέος Ελληνομνήμων Κ΄ (1926), p. 363–365.Wayback Machine Rubio y Lluch A., "Συμβολαί εις την ιστορίαν των Καταλωνίων εν Ελλάδι", Δελτίον της Ιστορικής και Εθνολογικής Εταιρείας της Ελλάδος Β΄(1883), p. 458–466. Lambros Spyridon ed., Έγγραφα αναφερόμενα εις την μεσαιωνικήν ιστορίαν των Αθηνών, Athens 1906. D' Olwer Nic., Les seigneurs Catalans d' Egine, τόμος εις μνήμην του Σπυρίδωνος Λάμπρου, Athens 1935. Koulikourdi Georgia, Αίγινα, 2 vols., Athens 1990. Μεσσίνας, Ηλίας, Οι Συναγωγές της Θεσσαλονίκης και της Βέροιας, Aθήνα 1997. . Messinas, Elias, The Synagogues of Greece: A Study of Synagogues in Macedonia and Thrace: With Architectural Drawings of all Synagogues of Greece. Seattle 2022. Μεσσίνας, Ηλίας, H Συναγωγή, Αθήνα: Εκδόσεις Ινφογνώμων 2022. Moutsopoulos Nikolaos, Η Παλιαχώρα της Αιγίνης. Ιστορική και μορφολογική εξέτασις των μνημείων, Athens 1962. Nikoloudis Nikolaos .:BiblioNet : Νικολούδης, Νικόλαος Γ., "Η Αίγινα κατά τον Μεσαίωνα και την Τουρκοκρατία", Βυζαντινός Δόμος 7(1993–94), pp:13–21. Pennas Charalambos .:BiblioNet : Πέννας, Χαράλαμπος, The Byzantine Aegina.:BiblioNet : Byzantine Aegina / Πέννας, Χαράλαμπος, Athens 2004. John N. Koumanoudes .:BiblioNet : Κουμανούδης, Ιωάννης Ν., Ανεμομυλικά ΙΙ, Αγκίστρι, Αίγινα, Αστυπάλαια, Λήμνος, Σαλαμίνα, Σπέτσες, Σύμη, Χίος και Ψαρά.:BiblioNet : Ανεμομυλικά ΙΙ / Κουμανούδης, Ιωάννης Ν., Τεχνικό Επιμελητήριο Ελλάδας, 2010. External links The feud between Athens and Aegina The Municipality of Aegina – official website Site for visitors and tourists run by the Municipality of Aegina Richard Stillwell, ed. Princeton Encyclopedia of Classical Sites, 1976: "Aigina, Greece" Map of Ancient Greece (includes Aegina Island) AeginaGreece.com Tourist guide The Mosaic of Aegina Program Aegina Aegean islands Islands of Attica Landforms of Islands (regional unit) Municipalities of Attica Provinces of Greece Mediterranean port cities and towns in Greece Populated places in Islands (regional unit) Stato da Màr Volcanoes of Greece Members of the Delian League Saronic Gulf Capitals of Greek states
2766
https://en.wikipedia.org/wiki/Ames%20test
Ames test
The Ames test is a widely employed method that uses bacteria to test whether a given chemical can cause mutations in the DNA of the test organism. More formally, it is a biological assay to assess the mutagenic potential of chemical compounds. A positive test indicates that the chemical is mutagenic and therefore may act as a carcinogen, because cancer is often linked to mutation. The test serves as a quick and convenient assay to estimate the carcinogenic potential of a compound because standard carcinogen assays on mice and rats are time-consuming (taking two to three years to complete) and expensive. However, false-positives and false-negatives are known. The procedure was described in a series of papers in the early 1970s by Bruce Ames and his group at the University of California, Berkeley. General procedure The Ames test uses several strains of the bacterium Salmonella typhimurium that carry mutations in genes involved in histidine synthesis. These strains are auxotrophic mutants, i.e. they require histidine for growth, but cannot produce it. The method tests the capability of the tested substance in creating mutations that result in a return to a "prototrophic" state, so that the cells can grow on a histidine-free medium. The tester strains are specially constructed to detect either frameshift (e.g. strains TA-1537 and TA-1538) or point (e.g. strain TA-1531) mutations in the genes required to synthesize histidine, so that mutagens acting via different mechanisms may be identified. Some compounds are quite specific, causing reversions in just one or two strains. The tester strains also carry mutations in the genes responsible for lipopolysaccharide synthesis, making the cell wall of the bacteria more permeable, and in the excision repair system to make the test more sensitive. Larger organisms like mammals have metabolic processes that could potentially turn a chemical considered not mutagenic into one that is or one that is considered mutagenic into one that is not. Therefore, to more effectively test a chemical compound's mutagenicity in relation to larger organisms, rat liver enzymes can be added in an attempt to replicate the metabolic processes' effect on the compound being tested in the Ames Test. Rat liver extract is optionally added to simulate the effect of metabolism, as some compounds, like benzo[a]pyrene, are not mutagenic themselves but their metabolic products are. The bacteria are spread on an agar plate with small amount of histidine. This small amount of histidine in the growth medium allows the bacteria to grow for an initial time and have the opportunity to mutate. When the histidine is depleted only bacteria that have mutated to gain the ability to produce its own histidine will survive. The plate is incubated for 48 hours. The mutagenicity of a substance is proportional to the number of colonies observed. Ames test and carcinogens Mutagens identified via Ames test are also possible carcinogens, and early studies by Ames showed that 90% of known carcinogens may be identified via this test. Later studies however showed identification of 50–70% of known carcinogens. The test was used to identify a number of compounds previously used in commercial products as potential carcinogens. Examples include tris(2,3-dibromopropyl)phosphate, which was used as a flame retardant in plastic and textiles such as children's sleepwear, and furylfuramide which was used as an antibacterial additive in food in Japan in the 1960s and 1970s. Furylfuramide in fact had previously passed animal tests, but more vigorous tests after its identification in the Ames test showed it to be carcinogenic. Their positive tests resulted in those chemicals being withdrawn from use in consumer products. One interesting result from the Ames test is that the dose response curve using varying concentrations of the chemical is almost always linear, indicating that there is no threshold concentration for mutagenesis. It therefore suggests that, as with radiation, there may be no safe threshold for chemical mutagens or carcinogens. However, some have proposed that organisms could tolerate low levels of mutagens due to protective mechanisms such as DNA repair, and thus a threshold may exist for certain chemical mutagens. Bruce Ames himself argued against linear dose-response extrapolation from the high dose used in carcinogenesis tests in animal systems to the lower dose of chemicals normally encountered in human exposure, as the results may be false positives due to mitogenic response caused by the artificially high dose of chemicals used in such tests. He also cautioned against the "hysteria over tiny traces of chemicals that may or may not cause cancer", that "completely drives out the major risks you should be aware of". The Ames test is often used as one of the initial screens for potential drugs to weed out possible carcinogens, and it is one of the eight tests required under the Pesticide Act (USA) and one of the six tests required under the Toxic Substances Control Act (USA). Limitations Salmonella typhimurium is a prokaryote, therefore it is not a perfect model for humans. Rat liver S9 fraction is used to mimic the mammalian metabolic conditions so that the mutagenic potential of metabolites formed by a parent molecule in the hepatic system can be assessed; however, there are differences in metabolism between humans and rats that can affect the mutagenicity of the chemicals being tested. The test may therefore be improved by the use of human liver S9 fraction; its use was previously limited by its availability, but it is now available commercially and therefore may be more feasible. An adapted in vitro model has been made for eukaryotic cells, for example yeast. Mutagens identified in the Ames test need not necessarily be carcinogenic, and further tests are required for any potential carcinogen identified in the test. Drugs that contain the nitrate moiety sometimes come back positive for Ames when they are indeed safe. The nitrate compounds may generate nitric oxide, an important signal molecule that can give a false positive. Nitroglycerin is an example that gives a positive Ames yet is still used in treatment today. Nitrates in food however may be reduced by bacterial action to nitrites which are known to generate carcinogens by reacting with amines and amides. Long toxicology and outcome studies are needed with such compounds to disprove a positive Ames test. Fluctuation method The Ames test was initially developed using agar plates (the plate incorporation technique), as described above. Since that time, an alternative to performing the Ames test has been developed, which is known as the "fluctuation method". This technique is the same in concept as the agar-based method, with bacteria being added to a reaction mixture with a small amount of histidine, which allows the bacteria to grow and mutate, returning to synthesize their own histidine. By including a pH indicator, the frequency of mutation is counted in microplates as the number of wells which have changed color (caused by a drop in pH due to metabolic processes of reproducing bacteria). As with the traditional Ames test, the sample is compared to the natural background rate of reverse mutation in order to establish the genotoxicity of a substance. The fluctuation method is performed entirely in liquid culture and is scored by counting the number of wells that turn yellow from purple in 96-well or 384-well microplates. In the 96-well plate method the frequency of mutation is counted as the number of wells out of 96 which have changed color. The plates are incubated for up to five days, with mutated (yellow) colonies being counted each day and compared to the background rate of reverse mutation using established tables of significance to determine the significant differences between the background rate of mutation and that for the tested samples. In the more scaled-down 384-well plate microfluctuation method the frequency of mutation is counted as the number of wells out of 48 which have changed color after 2 days of incubation. A test sample is assayed across 6 dose levels with concurrent zero-dose (background) and positive controls which all fit into one 384-well plate. The assay is performed in triplicates to provide statistical robustness. It uses the recommended OECD Guideline 471 tester strains (histidine auxotrophs and tryptophan auxotrophs). The fluctuation method is comparable to the traditional pour plate method in terms of sensitivity and accuracy, however, it does have a number of advantages: it needs less test sample, it has a simple colorimetric endpoint, counting the number of positive wells out of possible 96 or 48 wells is much less time-consuming than counting individual colonies on an agar plate. Several commercial kits are available. Most kits have consumable components in a ready-to-use state, including lyophilized bacteria, and tests can be performed using multichannel pipettes. The fluctuation method also allows for testing higher volumes of aqueous samples (up to 75% v/v), increasing the sensitivity and extending its application to low-level environmental mutagens. References Further reading Applied genetics Laboratory techniques Toxicology tests
2935
https://en.wikipedia.org/wiki/Ans
Ans
Ans or ANS or variation, may refer to: Places Ans, Belgium, a municipality in Belgium Ans, Denmark, a village in Denmark Angus, Scotland, UK; a council area by its Chapman code Ainsdale railway station, England, UK (by station code ANS) Andahuaylas Airport, Peru (by IATA airport code ANS) People Ans (given name), a Dutch feminine given name Anna Nicole Smith, American model and actress Organizations Academy of Natural Sciences of Philadelphia, Pennsylvania, United States Astronomical Netherlands Satellite, a Dutch satellite American Name Society American Nuclear Society American Numismatic Society, formerly the American Numismatic and Archaeological Society ANS Group of Companies, a news organization in Azerbaijan , a Cambodian resistance group; see Coalition Government of Democratic Kampuchea Audubon Naturalist Society, an American environmental organization Chemistry and biology Adrenergic nervous system, adrenaline and noradrenaline neurotransmitters distribution in human body 8-Anilino-1-naphthalenesulfonic acid, a fluorescent chemical compound used as a molecular probe Anthocyanidin synthase, an enzyme in the leucocyanidin biosynthesis pathway Approximate number system, a hypothesized physiological basis for the sense of number Autonomic nervous system, part of the peripheral nervous system in the body Technology , an unofficial file extension for ANSI art Advanced Network and Services, a non-profit network service provider in the 1990s American National Standards, defined by the American National Standards Institute ans, a variable in MATLAB, created automatically when no output argument is specified, referring to the most recent answer ANS carriage control characters (or ASA control characters), for computer line printers Asymmetric numeral systems, coding in data compression Authoritative name server, a DNS server Artificial neural system, or Artificial neural network Air Navigation Services, as delivered by an Air Navigation Service Provider (ANSP) Music ANS (album), a box set from the British band Coil ANS synthesizer, a Russian photoelectric musical instrument Other uses Al Ansar FC, a Lebanese association football club Amman National School, in Amman, Jordan Ansvarlig selskap, a Norwegian personal responsibility company model Algemeen Nijmeegs Studentenblad, a Dutch student magazine See also AN (disambiguation) Answer (disambiguation), for which "Ans." may be an abbreviation
3090
https://en.wikipedia.org/wiki/Arithmetic%E2%80%93geometric%20mean
Arithmetic–geometric mean
In mathematics, the arithmetic–geometric mean of two positive real numbers and is the mutual limit of a sequence of arithmetic means and a sequence of geometric means: Begin the sequences with x and y: Then define the two interdependent sequences and as These two sequences converge to the same number, the arithmetic–geometric mean of and ; it is denoted by , or sometimes by or . The arithmetic–geometric mean is used in fast algorithms for exponential and trigonometric functions, as well as some mathematical constants, in particular, computing . The arithmetic–geometric mean can be extended to complex numbers and when the branches of the square root are allowed to be taken inconsistently, it is, in general, a multivalued function. Example To find the arithmetic–geometric mean of and , iterate as follows: The first five iterations give the following values: The number of digits in which and agree (underlined) approximately doubles with each iteration. The arithmetic–geometric mean of 24 and 6 is the common limit of these two sequences, which is approximately . History The first algorithm based on this sequence pair appeared in the works of Lagrange. Its properties were further analyzed by Gauss. Properties The geometric mean of two positive numbers is never bigger than the arithmetic mean (see inequality of arithmetic and geometric means). As a consequence, for , is an increasing sequence, is a decreasing sequence, and . These are strict inequalities if . is thus a number between the geometric and arithmetic mean of and ; it is also between and . If , then . There is an integral-form expression for : where is the complete elliptic integral of the first kind: Indeed, since the arithmetic–geometric process converges so quickly, it provides an efficient way to compute elliptic integrals via this formula. In engineering, it is used for instance in elliptic filter design. The arithmetic–geometric mean is connected to the Jacobi theta function by which upon setting gives Related concepts The reciprocal of the arithmetic–geometric mean of 1 and the square root of 2 is called Gauss's constant, after Carl Friedrich Gauss. In 1799, Gauss proved that where is the lemniscate constant. In 1941, (and hence ) was proven transcendental by Theodor Schneider. The set is algebraically independent over , but the set (where the prime denotes the derivative with respect to the second variable) is not algebraically independent over . In fact, The geometric–harmonic mean can be calculated by an analogous method, using sequences of geometric and harmonic means. One finds that . The arithmetic–harmonic mean can be similarly defined, but takes the same value as the geometric mean (see section "Calculation" there). The arithmetic–geometric mean can be used to compute – among others – logarithms, complete and incomplete elliptic integrals of the first and second kind, and Jacobi elliptic functions. Proof of existence From the inequality of arithmetic and geometric means we can conclude that: and thus that is, the sequence is nondecreasing. Furthermore, it is easy to see that it is also bounded above by the larger of and (which follows from the fact that both the arithmetic and geometric means of two numbers lie between them). Thus, by the monotone convergence theorem, the sequence is convergent, so there exists a such that: However, we can also see that: and so: Q.E.D. Proof of the integral-form expression This proof is given by Gauss. Let Changing the variable of integration to , where gives Thus, we have The last equality comes from observing that . Finally, we obtain the desired result Applications The number π For example, according to the Gauss–Legendre algorithm: where with and , which can be computed without loss of precision using Complete elliptic integral K(sinα) Taking and yields the AGM where is a complete elliptic integral of the first kind: That is to say that this quarter period may be efficiently computed through the AGM, Other applications Using this property of the AGM along with the ascending transformations of John Landen, Richard P. Brent suggested the first AGM algorithms for the fast evaluation of elementary transcendental functions (, , ). Subsequently, many authors went on to study the use of the AGM algorithms. See also Landen's transformation Gauss–Legendre algorithm Generalized mean References Notes Citations Sources Means Special functions Elliptic functions Articles containing proofs
3292
https://en.wikipedia.org/wiki/Brass
Brass
Brass is an alloy of copper (Cu) and zinc (Zn), in proportions which can be varied to achieve different colours and mechanical, electrical, and chemical properties, but copper typically has the larger proportion. In use since prehistoric times, it is a substitutional alloy: atoms of the two constituents may replace each other within the same crystal structure. Brass is similar to bronze, another copper alloy that uses tin instead of zinc. Both bronze and brass may include small proportions of a range of other elements including arsenic (As), lead (Pb), phosphorus (P), aluminium (Al), manganese (Mn), and silicon (Si). Historically, the distinction between the two alloys has been less consistent and clear, and increasingly museums use the more general term "copper alloy." Brass has long been a popular material for its bright gold-like appearance and is still used for drawer pulls and doorknobs. It has also been widely used to make sculpture and utensils because of its low melting point, high workability (both with hand tools and with modern turning and milling machines), durability, and electrical and thermal conductivity. Brasses with higher copper content are softer and more golden in colour; conversely those with less copper and thus more zinc are harder and more silvery in colour. Brass is still commonly used in applications where corrosion resistance and low friction are required, such as locks, hinges, gears, bearings, ammunition casings, zippers, plumbing, hose couplings, valves, and electrical plugs and sockets. It is used extensively for musical instruments such as horns and bells. The composition of brass, generally 66% copper and 34% zinc, makes it a favorable substitute for copper in costume jewelry and fashion jewelry, as it exhibits greater resistance to corrosion. Brass is not as hard as bronze, and so is not suitable for most weapons and tools nor marine uses, because the zinc reacts with minerals in salt water, leaving porous copper behind; Marine brass, with added tin, avoids this, as does bronze. Brass is often used in situations in which it is important that sparks not be struck, such as in fittings and tools used near flammable or explosive materials. Properties Brass is more malleable than bronze or zinc. The relatively low melting point of brass (, depending on composition) and its flow characteristics make it a relatively easy material to cast. By varying the proportions of copper and zinc, the properties of the brass can be changed, allowing hard and soft brasses. The density of brass is . Today, almost 90% of all brass alloys are recycled. Because brass is not ferromagnetic, ferrous scrap can be separated from it by passing the scrap near a powerful magnet. Brass scrap is melted and recast into billets that are extruded into the desired form and size. The general softness of brass means that it can often be machined without the use of cutting fluid, though there are exceptions to this. Aluminium makes brass stronger and more corrosion-resistant. Aluminium also causes a highly beneficial hard layer of aluminium oxide (Al2O3) to be formed on the surface that is thin, transparent, and self-healing. Tin has a similar effect and finds its use especially in seawater applications (naval brasses). Combinations of iron, aluminium, silicon, and manganese make brass wear- and tear-resistant. The addition of as little as 1% iron to a brass alloy will result in an alloy with a noticeable magnetic attraction. Brass will corrode in the presence of moisture, chlorides, acetates, ammonia, and certain acids. This often happens when the copper reacts with sulfur to form a brown and eventually black surface layer of copper sulfide which, if regularly exposed to slightly acidic water such as urban rainwater, can then oxidize in air to form a patina of green-blue copper carbonate. Depending on how the patina layer was formed, it may protect the underlying brass from further damage. Although copper and zinc have a large difference in electrical potential, the resulting brass alloy does not experience internalized galvanic corrosion because of the absence of a corrosive environment within the mixture. However, if brass is placed in contact with a more noble metal such as silver or gold in such an environment, the brass will corrode galvanically; conversely, if brass is in contact with a less-noble metal such as zinc or iron, the less noble metal will corrode and the brass will be protected. Lead content To enhance the machinability of brass, lead is often added in concentrations of about 2%. Since lead has a lower melting point than the other constituents of the brass, it tends to migrate towards the grain boundaries in the form of globules as it cools from casting. The pattern the globules form on the surface of the brass increases the available lead surface area which, in turn, affects the degree of leaching. In addition, cutting operations can smear the lead globules over the surface. These effects can lead to significant lead leaching from brasses of comparatively low lead content. In October 1999, the California State Attorney General sued 13 key manufacturers and distributors over lead content. In laboratory tests, state researchers found the average brass key, new or old, exceeded the California Proposition 65 limits by an average factor of 19, assuming handling twice a day. In April 2001 manufacturers agreed to reduce lead content to 1.5%, or face a requirement to warn consumers about lead content. Keys plated with other metals are not affected by the settlement, and may continue to use brass alloys with a higher percentage of lead content. Also in California, lead-free materials must be used for "each component that comes into contact with the wetted surface of pipes and pipe fittings, plumbing fittings and fixtures". On 1 January 2010, the maximum amount of lead in "lead-free brass" in California was reduced from 4% to 0.25% lead. Corrosion-resistant brass for harsh environments Dezincification-resistant (DZR or DR) brasses, sometimes referred to as CR (corrosion resistant) brasses, are used where there is a large corrosion risk and where normal brasses do not meet the requirements. Applications with high water temperatures, chlorides present or deviating water qualities (soft water) play a role. DZR-brass is excellent in water boiler systems. This brass alloy must be produced with great care, with special attention placed on a balanced composition and proper production temperatures and parameters to avoid long-term failures. An example of DZR brass is the C352 brass, with about 30% zinc, 61–63% copper, 1.7–2.8% lead, and 0.02–0.15% arsenic. The lead and arsenic significantly suppress the zinc loss. "Red brasses", a family of alloys with high copper proportion and generally less than 15% zinc, are more resistant to zinc loss. One of the metals called "red brass" is 85% copper, 5% tin, 5% lead, and 5% zinc. Copper alloy C23000, which is also known as "red brass", contains 84–86% copper, 0.05% each iron and lead, with the balance being zinc. Another such material is gunmetal, from the family of red brasses. Gunmetal alloys contain roughly 88% copper, 8-10% tin, and 2-4% zinc. Lead can be added for ease of machining or for bearing alloys. "Naval brass", for use in seawater, contains 40% zinc but also 1% tin. The tin addition suppresses zinc leaching. The NSF International requires brasses with more than 15% zinc, used in piping and plumbing fittings, to be dezincification-resistant. Use in musical instruments The high malleability and workability, relatively good resistance to corrosion, and traditionally attributed acoustic properties of brass, have made it the usual metal of choice for construction of musical instruments whose acoustic resonators consist of long, relatively narrow tubing, often folded or coiled for compactness; silver and its alloys, and even gold, have been used for the same reasons, but brass is the most economical choice. Collectively known as brass instruments, these include the trombone, tuba, trumpet, cornet, flugelhorn, baritone horn, euphonium, tenor horn, and French horn, and many other "horns", many in variously sized families, such as the saxhorns. Other wind instruments may be constructed of brass or other metals, and indeed most modern student-model flutes and piccolos are made of some variety of brass, usually a cupronickel alloy similar to nickel silver (also known as German silver). Clarinets, especially low clarinets such as the contrabass and subcontrabass, are sometimes made of metal because of limited supplies of the dense, fine-grained tropical hardwoods traditionally preferred for smaller woodwinds. For the same reason, some low clarinets, bassoons and contrabassoons feature a hybrid construction, with long, straight sections of wood, and curved joints, neck, and/or bell of metal. The use of metal also avoids the risks of exposing wooden instruments to changes in temperature or humidity, which can cause sudden cracking. Even though the saxophones and sarrusophones are classified as woodwind instruments, they are normally made of brass for similar reasons, and because their wide, conical bores and thin-walled bodies are more easily and efficiently made by forming sheet metal than by machining wood. The keywork of most modern woodwinds, including wooden-bodied instruments, is also usually made of an alloy such as nickel silver. Such alloys are stiffer and more durable than the brass used to construct the instrument bodies, but still workable with simple hand tools—a boon to quick repairs. The mouthpieces of both brass instruments and, less commonly, woodwind instruments are often made of brass among other metals as well. Next to the brass instruments, the most notable use of brass in music is in various percussion instruments, most notably cymbals, gongs, and orchestral (tubular) bells (large "church" bells are normally made of bronze). Small handbells and "jingle bells" are also commonly made of brass. The harmonica is a free reed aerophone, also often made from brass. In organ pipes of the reed family, brass strips (called tongues) are used as the reeds, which beat against the shallot (or beat "through" the shallot in the case of a "free" reed). Although not part of the brass section, snare drums are also sometimes made of brass. Some parts on electric guitars are also made from brass, especially inertia blocks on tremolo systems for its tonal properties, and for string nuts and saddles for both tonal properties and its low friction. Germicidal and antimicrobial applications The bactericidal properties of brass have been observed for centuries, particularly in marine environments where it prevents biofouling. Depending upon the type and concentration of pathogens and the medium they are in, brass kills these microorganisms within a few minutes to hours of contact. A large number of independent studies confirm this antimicrobial effect, even against antibiotic-resistant bacteria such as MRSA and VRSA. The mechanisms of antimicrobial action by copper and its alloys, including brass, are a subject of intense and ongoing investigation. Season cracking Brass is susceptible to stress corrosion cracking, especially from ammonia or substances containing or releasing ammonia. The problem is sometimes known as season cracking after it was first discovered in brass cartridges used for rifle ammunition during the 1920s in the British Indian Army. The problem was caused by high residual stresses from cold forming of the cases during manufacture, together with chemical attack from traces of ammonia in the atmosphere. The cartridges were stored in stables and the ammonia concentration rose during the hot summer months, thus initiating brittle cracks. The problem was resolved by annealing the cases, and storing the cartridges elsewhere. Types Other phases than α, β and γ are ε, a hexagonal intermetallic CuZn3, and η, a solid solution of copper in zinc. History Although forms of brass have been in use since prehistory, its true nature as a copper-zinc alloy was not understood until the post-medieval period because the zinc vapor which reacted with copper to make brass was not recognized as a metal. The King James Bible makes many references to "brass" to translate "nechosheth" (bronze or copper) from Hebrew to English. The earliest brasses may have been natural alloys made by smelting zinc-rich copper ores. By the Roman period brass was being deliberately produced from metallic copper and zinc minerals using the cementation process, the product of which was calamine brass, and variations on this method continued until the mid-19th century. It was eventually replaced by speltering, the direct alloying of copper and zinc metal which was introduced to Europe in the 16th century. Brass has sometimes historically been referred to as "yellow copper". Early copper-zinc alloys In West Asia and the Eastern Mediterranean early copper-zinc alloys are now known in small numbers from a number of 3rd millennium BC sites in the Aegean, Iraq, the United Arab Emirates, Kalmykia, Turkmenistan and Georgia and from 2nd millennium BC sites in West India, Uzbekistan, Iran, Syria, Iraq and Canaan. Isolated examples of copper-zinc alloys are known in China from the 1st century AD, long after bronze was widely used. The compositions of these early "brass" objects are highly variable and most have zinc contents of between 5% and 15% wt which is lower than in brass produced by cementation. These may be "natural alloys" manufactured by smelting zinc rich copper ores in redox conditions. Many have similar tin contents to contemporary bronze artefacts and it is possible that some copper-zinc alloys were accidental and perhaps not even distinguished from copper. However the large number of copper-zinc alloys now known suggests that at least some were deliberately manufactured and many have zinc contents of more than 12% wt which would have resulted in a distinctive golden colour. By the 8th–7th century BC Assyrian cuneiform tablets mention the exploitation of the "copper of the mountains" and this may refer to "natural" brass. "Oreikhalkon" (mountain copper), the Ancient Greek translation of this term, was later adapted to the Latin aurichalcum meaning "golden copper" which became the standard term for brass. In the 4th century BC Plato knew orichalkos as rare and nearly as valuable as gold and Pliny describes how aurichalcum had come from Cypriot ore deposits which had been exhausted by the 1st century AD. X-ray fluorescence analysis of 39 orichalcum ingots recovered from a 2,600-year-old shipwreck off Sicily found them to be an alloy made with 75–80% copper, 15–20% zinc and small percentages of nickel, lead and iron. Roman world During the later part of first millennium BC the use of brass spread across a wide geographical area from Britain and Spain in the west to Iran, and India in the east. This seems to have been encouraged by exports and influence from the Middle East and eastern Mediterranean where deliberate production of brass from metallic copper and zinc ores had been introduced. The 4th century BC writer Theopompus, quoted by Strabo, describes how heating earth from Andeira in Turkey produced "droplets of false silver", probably metallic zinc, which could be used to turn copper into oreichalkos. In the 1st century BC the Greek Dioscorides seems to have recognized a link between zinc minerals and brass describing how Cadmia (zinc oxide) was found on the walls of furnaces used to heat either zinc ore or copper and explaining that it can then be used to make brass. By the first century BC brass was available in sufficient supply to use as coinage in Phrygia and Bithynia, and after the Augustan currency reform of 23 BC it was also used to make Roman dupondii and sestertii. The uniform use of brass for coinage and military equipment across the Roman world may indicate a degree of state involvement in the industry, and brass even seems to have been deliberately boycotted by Jewish communities in Palestine because of its association with Roman authority. Brass was produced by the cementation process where copper and zinc ore are heated together until zinc vapor is produced which reacts with the copper. There is good archaeological evidence for this process and crucibles used to produce brass by cementation have been found on Roman period sites including Xanten and Nidda in Germany, Lyon in France and at a number of sites in Britain. They vary in size from tiny acorn sized to large amphorae like vessels but all have elevated levels of zinc on the interior and are lidded. They show no signs of slag or metal prills suggesting that zinc minerals were heated to produce zinc vapor which reacted with metallic copper in a solid state reaction. The fabric of these crucibles is porous, probably designed to prevent a buildup of pressure, and many have small holes in the lids which may be designed to release pressure or to add additional zinc minerals near the end of the process. Dioscorides mentioned that zinc minerals were used for both the working and finishing of brass, perhaps suggesting secondary additions. Brass made during the early Roman period seems to have varied between 20% and 28% wt zinc. The high content of zinc in coinage and brass objects declined after the first century AD and it has been suggested that this reflects zinc loss during recycling and thus an interruption in the production of new brass. However it is now thought this was probably a deliberate change in composition and overall the use of brass increases over this period making up around 40% of all copper alloys used in the Roman world by the 4th century AD. Medieval period Little is known about the production of brass during the centuries immediately after the collapse of the Roman Empire. Disruption in the trade of tin for bronze from Western Europe may have contributed to the increasing popularity of brass in the east and by the 6th–7th centuries AD over 90% of copper alloy artefacts from Egypt were made of brass. However other alloys such as low tin bronze were also used and they vary depending on local cultural attitudes, the purpose of the metal and access to zinc, especially between the Islamic and Byzantine world. Conversely the use of true brass seems to have declined in Western Europe during this period in favor of gunmetals and other mixed alloys but by about 1000 brass artefacts are found in Scandinavian graves in Scotland, brass was being used in the manufacture of coins in Northumbria and there is archaeological and historical evidence for the production of calamine brass in Germany and the Low Countries, areas rich in calamine ore. These places would remain important centres of brass making throughout the Middle Ages period, especially Dinant. Brass objects are still collectively known as dinanderie in French. The baptismal font at St Bartholomew's Church, Liège in modern Belgium (before 1117) is an outstanding masterpiece of Romanesque brass casting, though also often described as bronze. The metal of the early 12th-century Gloucester Candlestick is unusual even by medieval standards in being a mixture of copper, zinc, tin, lead, nickel, iron, antimony and arsenic with an unusually large amount of silver, ranging from 22.5% in the base to 5.76% in the pan below the candle. The proportions of this mixture may suggest that the candlestick was made from a hoard of old coins, probably Late Roman. Latten is a term for medieval alloys of uncertain and often variable composition often covering decorative borders and similar objects cut from sheet metal, whether of brass or bronze. Especially in Tibetan art, analysis of some objects shows very different compositions from different ends of a large piece. Aquamaniles were typically made in brass in both the European and Islamic worlds. The cementation process continued to be used but literary sources from both Europe and the Islamic world seem to describe variants of a higher temperature liquid process which took place in open-topped crucibles. Islamic cementation seems to have used zinc oxide known as tutiya or tutty rather than zinc ores for brass-making, resulting in a metal with lower iron impurities. A number of Islamic writers and the 13th century Italian Marco Polo describe how this was obtained by sublimation from zinc ores and condensed onto clay or iron bars, archaeological examples of which have been identified at Kush in Iran. It could then be used for brass making or medicinal purposes. In 10th century Yemen al-Hamdani described how spreading al-iglimiya, probably zinc oxide, onto the surface of molten copper produced tutiya vapor which then reacted with the metal. The 13th century Iranian writer al-Kashani describes a more complex process whereby tutiya was mixed with raisins and gently roasted before being added to the surface of the molten metal. A temporary lid was added at this point presumably to minimize the escape of zinc vapor. In Europe a similar liquid process in open-topped crucibles took place which was probably less efficient than the Roman process and the use of the term tutty by Albertus Magnus in the 13th century suggests influence from Islamic technology. The 12th century German monk Theophilus described how preheated crucibles were one sixth filled with powdered calamine and charcoal then topped up with copper and charcoal before being melted, stirred then filled again. The final product was cast, then again melted with calamine. It has been suggested that this second melting may have taken place at a lower temperature to allow more zinc to be absorbed. Albertus Magnus noted that the "power" of both calamine and tutty could evaporate and described how the addition of powdered glass could create a film to bind it to the metal. German brass making crucibles are known from Dortmund dating to the 10th century AD and from Soest and Schwerte in Westphalia dating to around the 13th century confirm Theophilus' account, as they are open-topped, although ceramic discs from Soest may have served as loose lids which may have been used to reduce zinc evaporation, and have slag on the interior resulting from a liquid process. Africa Some of the most famous objects in African art are the lost wax castings of West Africa, mostly from what is now Nigeria, produced first by the Kingdom of Ife and then the Benin Empire. Though normally described as "bronzes", the Benin Bronzes, now mostly in the British Museum and other Western collections, and the large portrait heads such as the Bronze Head from Ife of "heavily leaded zinc-brass" and the Bronze Head of Queen Idia, both also British Museum, are better described as brass, though of variable compositions. Work in brass or bronze continued to be important in Benin art and other West African traditions such as Akan goldweights, where the metal was regarded as a more valuable material than in Europe. Renaissance and post-medieval Europe The Renaissance saw important changes to both the theory and practice of brassmaking in Europe. By the 15th century there is evidence for the renewed use of lidded cementation crucibles at Zwickau in Germany. These large crucibles were capable of producing c.20 kg of brass. There are traces of slag and pieces of metal on the interior. Their irregular composition suggests that this was a lower temperature, not entirely liquid, process. The crucible lids had small holes which were blocked with clay plugs near the end of the process presumably to maximize zinc absorption in the final stages. Triangular crucibles were then used to melt the brass for casting. 16th-century technical writers such as Biringuccio, Ercker and Agricola described a variety of cementation brass making techniques and came closer to understanding the true nature of the process noting that copper became heavier as it changed to brass and that it became more golden as additional calamine was added. Zinc metal was also becoming more commonplace. By 1513 metallic zinc ingots from India and China were arriving in London and pellets of zinc condensed in furnace flues at the Rammelsberg in Germany were exploited for cementation brass making from around 1550. Eventually it was discovered that metallic zinc could be alloyed with copper to make brass, a process known as speltering, and by 1657 the German chemist Johann Glauber had recognized that calamine was "nothing else but unmeltable zinc" and that zinc was a "half ripe metal". However some earlier high zinc, low iron brasses such as the 1530 Wightman brass memorial plaque from England may have been made by alloying copper with zinc and include traces of cadmium similar to those found in some zinc ingots from China. However, the cementation process was not abandoned, and as late as the early 19th century there are descriptions of solid-state cementation in a domed furnace at around 900–950 °C and lasting up to 10 hours. The European brass industry continued to flourish into the post medieval period buoyed by innovations such as the 16th century introduction of water powered hammers for the production of wares such as pots. By 1559 the Germany city of Aachen alone was capable of producing 300,000 cwt of brass per year. After several false starts during the 16th and 17th centuries the brass industry was also established in England taking advantage of abundant supplies of cheap copper smelted in the new coal fired reverberatory furnace. In 1723 Bristol brass maker Nehemiah Champion patented the use of granulated copper, produced by pouring molten metal into cold water. This increased the surface area of the copper helping it react and zinc contents of up to 33% wt were reported using this new technique. In 1738 Nehemiah's son William Champion patented a technique for the first industrial scale distillation of metallic zinc known as distillation per descencum or "the English process". This local zinc was used in speltering and allowed greater control over the zinc content of brass and the production of high-zinc copper alloys which would have been difficult or impossible to produce using cementation, for use in expensive objects such as scientific instruments, clocks, brass buttons and costume jewelry. However Champion continued to use the cheaper calamine cementation method to produce lower-zinc brass and the archaeological remains of bee-hive shaped cementation furnaces have been identified at his works at Warmley. By the mid-to-late 18th century developments in cheaper zinc distillation such as John-Jaques Dony's horizontal furnaces in Belgium and the reduction of tariffs on zinc as well as demand for corrosion-resistant high zinc alloys increased the popularity of speltering and as a result cementation was largely abandoned by the mid-19th century. See also Brass bed Brass rubbing List of copper alloys Citations General references Bayley, J. (1990). "The Production of Brass in Antiquity with Particular Reference to Roman Britain". In Craddock, P. T. (ed.). 2000 Years of Zinc and Brass. London: British Museum. Craddock, P. T. and Eckstein, K (2003). "Production of Brass in Antiquity by Direct Reduction". In Craddock, P. T. and Lang, J. (eds.). Mining and Metal Production Through the Ages. London: British Museum. Day, J. (1990). "Brass and Zinc in Europe from the Middle Ages until the 19th century". In Craddock, P. T. (ed.). 2000 Years of Zinc and Brass. London: British Museum. Day, J. (1991). "Copper, Zinc and Brass Production". In Day, J. and Tylecote, R. F. (eds.). The Industrial Revolution in Metals. London: The Institute of Metals. Rehren, T. and Martinon Torres, M. (2008) "Naturam ars imitate: European brassmaking between craft and science". In Martinon-Torres, M. and Rehren, T. (eds.). Archaeology, History and Science: Integrating Approaches to Ancient Material. Left Coast Press. External links Copper alloys History of metallurgy Zinc alloys
3533
https://en.wikipedia.org/wiki/Government%20of%20Belarus
Government of Belarus
The Government of the Republic of Belarus (), which consists of the Council of Ministers of the Republic of Belarus (), is the executive branch of state power in Belarus, and is appointed by the President of Belarus. The head of the Government is the President of Belarus, who manages the main agenda of the government and direct the ministers. The National Assembly of Belarus is the continuation of the Supreme Soviet of the BSSR and acts as the functioning parliament for Belarus. Council of Ministers Below are the 30 members of the Council of Ministers as of 19 August 2020, as well as the head of the presidential administration and the chairmen of the State Committees, who are not technically ministers but are included in the Council of Ministers. Offices which are not technically counted as ministerial posts are italicized. The prime minister, the first deputy prime minister(s), the deputy prime minister(s), the ministers of economy, finance, and foreign affairs, the head of the presidential administration, and the chairman of the State Control Committee together form the Presidium of the Council of Ministers. These officials are highlighted in yellow. The incumbent government resigned en masse on 17 August 2020. A new government was formed on 19 August 2020, consisting of mostly the same people. Composition As of 18 October 2021: See also Supreme Soviet of Belarus, the preceding supreme state power in Belarus National Assembly of Belarus, the current supreme state power in Belarus References Belarus Belarus
3741
https://en.wikipedia.org/wiki/Benny%20Andersson
Benny Andersson
Göran Bror Benny Andersson (; born 16 December 1946) is a Swedish musician, singer, composer and producer best known as a member of the musical group ABBA and co-composer of the musicals Chess, Kristina från Duvemåla, and Mamma Mia! For the 2008 film version of Mamma Mia! and its 2018 sequel, Mamma Mia! Here We Go Again, he worked also as an executive producer. Since 2001, he has been active with his own band Benny Anderssons orkester. Early life Göran Bror Benny Andersson was born 16 December 1946 in the Vasastan district of Stockholm to civil engineer Gösta Andersson (1912–1973) and his wife Laila (1920–1971). His sister Eva-Lis Andersson followed in 1948. Andersson's musical background comes from his father and grandfather (Efraim); they both enjoyed playing the accordion, and at six, Benny got his own. His father and grandfather taught him Swedish folk music, traditional music, and schlager. The first records Benny bought were "Du Bist Musik" by Italian schlager singer Caterina Valente and Elvis Presley's "Jailhouse Rock". He was especially impressed by the flip side, "Treat Me Nice", as this featured a piano. This variety of different kinds of music influenced him through the years. At ten, Andersson got his own piano and taught himself to play. He left school at 15 and began to perform at youth clubs. This is when he met his first girlfriend Christina Grönvall, with whom he had two children: Peter (born 1963) and Heléne (born 1965). In early 1964, Benny and Christina joined "Elverkets Spelmanslag" ("The Electricity Board Folk Music Group"); the name was a punning reference to their electric instruments. Their repertoire consisted mainly of instrumentals, including "Baby Elephant Walk"; he also wrote his first songs. Career Hep Stars (1964–1969) In October 1964 he joined the Hep Stars as keyboardist and they made a breakthrough in March 1965 with their hit "Cadillac", eventually becoming the most celebrated of the Swedish 1960s pop bands. Andersson secured his place as the band's keyboardist and musical driving force as well as a teen idol. The band performed mostly covers of international hits, but Andersson soon started writing his own material, and gave the band the classic hits "No Response", "Sunny Girl", "Wedding", "Consolation", "It's Nice To Be Back" and "She Will Love You," among others. Before ABBA (1969–1972) Andersson met Björn Ulvaeus in June 1966, and the two men started writing songs together, their first being "Isn't It Easy To Say", eventually recorded by the Hep Stars. He also had a fruitful songwriting collaboration with Lasse Berghagen, with whom he wrote several songs and submitted "Hej, Clown" for the 1969 Melodifestivalen – the Swedish Eurovision Song Festival finals. The song finished in second place. During this contest, he met vocalist Anni-Frid Lyngstad, and they soon became a couple. Around the same time, his songwriting companion Ulvaeus met vocalist Agnetha Fältskog. The personal relationships and Andersson and Ulvaeus' songwriting collaboration led quite naturally to the very close cooperation which the four friends had during the following years. Benny and Björn scored their first hits as songwriters in the spring of 1969: "Ljuva sextital" (a hit with Brita Borg) and "Speleman" (a hit for the Hep Stars). As the two couples began supporting each other during recording sessions, the sound of the women's voices convinced the songwriters to model their 'group' on various MOR acts such as Blue Mink, Middle of the Road and Sweet. Thus, ABBA came to life. ABBA (1972–1982) The group's breakthrough came with winning the Eurovision Song Contest for Sweden with "Waterloo" on 6 April 1974. During the next eight years, Andersson (together with Ulvaeus) wrote music for and produced eight studio albums with ABBA. The group achieved great success globally and scored a chain of No. 1 hits. With ABBA, Benny sang lead vocal on only one song - "Suzy-Hang-Around", from the Waterloo album. After ABBA: Chess, Kristina and Mamma Mia! (1983–present) After ABBA, Andersson continued writing music with Ulvaeus. Their first project was the stage musical Chess, written with Tim Rice. The Chess concept album – with vocals by Elaine Paige, Barbara Dickson, Murray Head and Swedes Tommy Körberg and Björn Skifs – was released in October 1984, selling two million copies worldwide. The Paige/Dickson duet "I Know Him So Well" became a major UK No. 1 hit, and Murray Head's "One Night in Bangkok" gave Andersson/Ulvaeus a US No. 3 hit. Chess was staged in London's West End Prince Edward Theatre in May 1986 and received mixed to positive reviews, running for about three years. A revised staging on Broadway in April 1988 received poor reviews, running for two months. In 1985, Andersson produced and released an album with brother and sister Anders and Karin Glenmark, featuring new songs by Andersson/Ulvaeus. The duo named themselves Gemini, and a second album with more music by Björn and Benny was released in April 1987, containing the big hit "Mio My Mio"; also to be found on the soundtrack to the film Mio in the Land of Faraway, for which Andersson co-produced the music. In 1987, Andersson released his first solo album Klinga Mina Klockor ("Chime, My Bells"). All the music was written by and performed by himself on accordion, backed by the Orsa Spelmän (Orsa Folk Musicians) on fiddles. A second solo album followed: November 1989. In 1990, Andersson scored a Swedish No. 1 hit with "Lassie", sung by female cabaret group Ainbusk, for whom he also wrote the Svensktoppen hits "Älska Mig" and "Drömmarnas Golv". He decided to produce an album with Josefin Nilsson from this quartet, resulting in the 1993 English-language album Shapes, featuring ten new Andersson/Ulvaeus compositions. In 1992, he wrote the introduction melody for the European football championship, which was organised by Sweden that year. From the late 1980s, Andersson had worked on an idea for an epic Swedish language musical based on his affection for traditional folk music, and in October 1995, Kristina från Duvemåla premiered in Sweden. The musical was based on The Emigrants novels by Swedish writer Vilhelm Moberg. The musical ran successfully for almost five years, before closing in June 1999. An English-language version, simply titled Kristina, was staged in concert at Carnegie Hall in New York City for two nights in September 2009, yielding a live recording, and at the Royal Albert Hall for one night in April 2010. Andersson's next project was Mamma Mia!, a musical built around 24 of ABBA's songs, which has become a worldwide box-office blockbuster, with versions in several languages being played in many countries, including the UK (West End premiere in April 1999), Canada (Toronto premiere in 2000), the USA (Broadway premiere in 2001), and Sweden (Swedish language premiere in 2005). For the 2004 semi-final of the Eurovision Song Contest, staged in Istanbul thirty years after ABBA had won the contest in Brighton, Benny appeared briefly in a special comedy video made for the interval act, entitled "Our Last Video". Each of the four members of the group appeared briefly in cameo roles, as did others such as Cher and Rik Mayall. The video was not included in the official DVD release of the Eurovision Contest, but was issued as a separate DVD release. It was billed as the first time the four had worked together since the group split; however, Frida's appearance was filmed separately. A film version of Mamma Mia! premiered on 18 July 2008. In April/May 2007, Andersson worked on the film soundtrack, re-recording the ABBA songs with musicians from the original ABBA recording sessions. Mamma Mia! The Movie has become the most successful film musical of all time, and the biggest-selling DVD ever in the UK. Benny Anderssons Orkester (2001–present) Andersson currently performs with his own band of 16 musicians, Benny Anderssons Orkester ("Benny Andersson's orchestra", BAO), with fellow Swedes Helen Sjöholm (of Kristina from Duvemåla) and Tommy Körberg (of Chess), with lyrics to new material sometimes written by Björn Ulvaeus. BAO has released five albums to huge success in Sweden, all containing hit singles. In 2009 BAO achieved a new record in Sweden on the Svensktoppen chart by staying there for 243 weeks with the song "Du är min man" ("You Are My Man"), sung by Sjöholm. New compositions (1984–present) Andersson composes primarily for his band BAO, with vocalists Sjöholm and Körberg, but he keeps his older material alive by re-visiting it, as in Mamma Mia! and the Swedish version of Chess. For a compilation album of the Glenmark duo Gemini, Andersson had Björn Ulvaeus write new Swedish lyrics for the re-recording of two songs from 1984 and 1987. Andersson and Ulvaeus have continuously been writing new material; most recently the two wrote seven songs for Andersson's BAO 2011 album O klang och jubeltid, performed as usual by vocalists Sjöholm, Körberg and Kalle Moraeus. In July 2009, BAO, now named "The Benny Andersson Band", released their first international record, the album The Story of a Heart. It was a compilation of 14 tracks from Andersson's five Swedish-language releases between 1987 and 2007, including five songs now recorded with lyrics by Ulvaeus in English, and the new title song premiered on BBC2's Ken Bruce Show. A Swedish-language version of the title track, "Sommaren Du Fick" ("The Summer You Got"), was released as a single in Sweden prior to the English version, with vocals by Helen Sjöholm. In the spring of 2009, Andersson also released a single recorded by the staff at his privately owned Stockholm hotel Hotel Rival, titled "2nd Best to None", accompanied by a video showing the staff at work. In 2008, Andersson and Ulvaeus wrote a song for Swedish singer Sissela Kyle, titled "Jag vill bli gammal" ("I Wanna Grow Old"), for her Stockholm stage show "Your Days Are Numbered", which was never recorded and released but did get a TV performance. Ulvaeus also contributed lyrics to ABBA's 1976 instrumental track "Arrival" for Sarah Brightman's cover version recorded for her 2008 album A Winter Symphony. New English lyrics have also been written for Andersson's 1999 song "Innan Gryningen" (then also named "Millennium Hymn"), with the new title "The Silence of the Dawn" for Barbara Dickson was performed live, but not yet recorded and released. In 2007, they wrote the new song "Han som har vunnit allt" ("He Who's Won It All") for actor and singer Anders Ekborg. Ulvaeus wrote English lyrics for two older songs from Andersson's solo albums I Walk with You Mama ("Stockholm by Night", 1989) and After the Rain ("Efter regnet", 1987) for opera singer Anne Sofie Von Otter, for her Andersson tribute album I Let the Music Speak. Barbara Dickson recorded an Ulvaeus and Andersson song called "The Day The Wall Came Tumbling Down"; the song eventually was released by Australian Mamma Mia! musical star Anne Wood's album of ABBA covers, Divine Discontent. As of October 2012, Ulvaeus had mentioned writing new material with Andersson for a BAO Christmas release (also mentioned as a BAO box), and Andersson is busy writing music for a Swedish language obscure musical, Hjälp Sökes ("Help is Wanted") together with Kristina Lugn and Lars Rudolfsson, premiering 8 February 2013. Andersson has also written music for a documentary film about Olof Palme, re-recording the track "Sorgmarch" from his last album as a theme throughout the film. The song "Kärlekens Tid", recorded 2004 by Helen Sjöholm with BAO, has also been performed in concert in English (lyrics by Ylva Eggehorn) by opera baritone Bryn Terfel. On 15 April 2013, it was officially announced by the EBU and the SVT that Andersson, along with Ulvaeus and the late Stockholm based DJ and record producer Avicii, had composed the anthem for the 2013 Eurovision Song Contest. The song was performed for the first time in the Final on 18 May. A new album of Andersson compositions presented in a choral style was released on 18 September 2015. "Kärlekens Tid" was produced in Andersson's Mono Music studio, under the direction of choirmaster Gustaf Sjökvist, who died before the album's release. Gustaf Sjökvists Choir, conducted by Cecilia Rydinger Alin, performed two concerts at Skeppsholmen on 20 September, featuring Benny Andersson on piano. The album includes songs in Swedish and English from a range of Andersson's projects, such as Chess, Kristina and BAO. In November 2018, Deutsche Grammophon released Piano, a collection of ABBA tunes, tunes from Chess and original compositions all played by Andersson on solo piano. Andersson reunited with ABBA in 2018. On 2 September 2021, via YouTube livestream, ABBA announced their upcoming virtual concert residency "ABBA Voyage", as well as the imminent release of an eponymous album, recorded between 2017 and 2021. The new record, their first studio album in 40 years, features ten tracks, including "I Still Have Faith In You" and "Don't Shut Me Down", which also were first shown in the aforementioned livestream event and released as a double A-side single. On 5 November 2021, the Voyage album was released worldwide. On 27 May 2022, ABBA Voyage opened in a purpose-built venue named the ABBA Arena at the Queen Elizabeth Olympic Park in London. Film music Andersson has written music to several films for screen and television; the first attempt in the early 1970s for the Swedish erotica movie The Seduction of Inga; the film was not a success, but the 'Björn & Benny' single "She's My Kind of Girl" surprised the composers by being released in Japan and becoming a Top 10 hit (the song renamed in Japan as "The Little Girl of the Cold Wind"). In 1987, Andersson wrote music and co-produced the soundtrack with Anders Eljas for the film Mio in the Land of Faraway, based on Swedish author Astrid Lindgrens Mio, My Son. The title song became a huge hit in Sweden for Gemini. In 2000, he wrote the music for fellow Swede (no relation) Roy Andersson's film Songs from the Second Floor (the music later re-recorded, featuring new lyrics, with BAO! with vocals by Helen Sjöholm). He also wrote the theme for Roy Andersson's next film, You, the Living, from 2007. Andersson worked on the film adaptation of Mamma Mia!. He also wrote the film score for the 2012 documentary Palme about Swedish prime minister Olof Palme. He later won a Guldbagge Award for Best Original Score, for that film at the 48th Guldbagge Awards. Awards Together with Ulvaeus, Andersson was nominated for a Drama Desk Award in a category "Outstanding Music" (for the musical Chess), and for a Tony Award in a category "Best Orchestrations" (for musical Mamma Mia!). Original cast recordings of both musicals were nominated for a Grammy Award. Andersson/Ulvaeus also won a Touring Broadway Award for the musical "Mamma Mia" (best score). During his post-ABBA career Andersson won four Swedish Grammis awards, and together with Ulvaeus received the "Special International" Ivor Novello award from 'The British Academy of Composers and Songwriters', twice "The Music Export Prize" from the Swedish Ministry of Industry and Trade (2008), as well as the "Lifetime Achievement" award from the Swedish Music Publishers Association (SMFF). In 2002, Andersson was given an honorary professorship by the Swedish Government for his "ability to create high-class music reaching people around the world". In 2007, he was elected a member of Royal Swedish Academy of Music, and in 2008 received an Honorary Doctorate from the Stockholm University Faculty of Humanities for contributing importantly both to the preservation and the growth of the Swedish folk music tradition. On 15 March 2010, Andersson appeared on stage in New York with former wife Anni-Frid Lyngstad to accept ABBA's award of induction into the Rock and Roll Hall of Fame. During his acceptance speech he reflected on the important influence of traditional European music and the melancholy of the Swedish soul on ABBA's brand of pop music. "If you live in a country like Sweden, with five, six months of snow, and the sun disappears totally for like two months, that would be reflected in the work of artists," he said. "It's definitely in the Swedish folk music, you can hear it in the Russian folk songs, you can hear in the music from Jean Sibelius or Edvard Grieg from Norway, you can see it in the eyes of Greta Garbo and you can hear it in the voice of Jussi Björling. And you can hear in the sound of Frida and Agnetha on some of our songs too." In 2012, he received an Honorary Doctorate of Philosophy from the Luleå Tekniska Universitet Faculty of Humanities and Social Sciences. Andersson won the Swedish "Guldbaggen" award in 2012 as composer of the music for the film "Palme". For his album Piano he received the Opus Klassik award in 2018. Personal life Andersson was engaged to Christina Grönvall. In 1963, they had a son, Peter, and in 1965, a daughter, Heléne. Both children were given their mother's surname. The couple split in 1966 and Christina kept custody of the children as Andersson was then at the peak of his Hep Stars' success. In the 1990s, Peter formed One More Time, a group that enjoyed European success with the ABBA-like "Highland" and, later, as Sweden's entry to the Eurovision Song Contest 1996. Andersson was in a relationship with Anni-Frid Lyngstad of ABBA for about 11 years, from 1969 till 1980. They married on 6 October 1978, but separated on 26 November 1980 and divorced in 1981. He married Swedish TV presenter Mona Nörklit in 1981 and had a son, Ludvig (born January 1982), who has since followed in his father's footsteps in forming his own band. Ludvig is also one of the producers of the concert residency ABBA Voyage. Andersson was an alcoholic through much of his adult life. He has remained a teetotaler since 2001. Andersson did not disclose the extent of his substance abuse problems until a 2011 interview, at which point he had maintained nearly a decade of sobriety. Discography Also appears on Beginner's Guide to Scandinavia (3CD, Nascente 2011) See also List of Swedes in music References External links Deutsche Grammophon page on Benny Andersson BBC Radio 3 New Generation Artists page icethesite (Benny Andersson and Björn Ulvaeus news site) The Hep Stars International Official website (Benny before ABBA) Chronology: Benny Andersson Official ABBA site 1946 births Living people Musicians from Stockholm ABBA members Members of the Royal Swedish Academy of Music Swedish accordionists Swedish keyboardists Swedish pianists Swedish male songwriters 20th-century Swedish male singers Swedish musical theatre composers Swedish film score composers Georges Delerue Award winners Swedish pop pianists Eurodisco musicians Ivor Novello Award winners Swedish feminists Male feminists Feminist musicians Eurovision Song Contest winners Best Original Score Guldbagge Award winners 21st-century Swedish male singers Acoustic guitarists Swedish organists Male organists 21st-century accordionists Male pianists 21st-century organists Male film score composers Deutsche Grammophon artists
3931
https://en.wikipedia.org/wiki/Binary%20relation
Binary relation
In mathematics, a binary relation associates elements of one set, called the domain, with elements of another set, called the codomain. A binary relation over sets and is a new set of ordered pairs consisting of elements in and in . It is a generalization of the more widely understood idea of a unary function. It encodes the common concept of relation: an element is related to an element , if and only if the pair belongs to the set of ordered pairs that defines the binary relation. A binary relation is the most studied special case of an -ary relation over sets , which is a subset of the Cartesian product An example of a binary relation is the "divides" relation over the set of prime numbers and the set of integers , in which each prime is related to each integer that is a multiple of , but not to an integer that is not a multiple of . In this relation, for instance, the prime number 2 is related to numbers such as −4, 0, 6, 10, but not to 1 or 9, just as the prime number 3 is related to 0, 6, and 9, but not to 4 or 13. Binary relations are used in many branches of mathematics to model a wide variety of concepts. These include, among others: the "is greater than", "is equal to", and "divides" relations in arithmetic; the "is congruent to" relation in geometry; the "is adjacent to" relation in graph theory; the "is orthogonal to" relation in linear algebra. A function may be defined as a special kind of binary relation. Binary relations are also heavily used in computer science. A binary relation over sets and is an element of the power set of Since the latter set is ordered by inclusion (⊆), each relation has a place in the lattice of subsets of A binary relation is called a homogeneous relation when X = Y. A binary relation is also called a heterogeneous relation when it is not necessary that X = Y. Since relations are sets, they can be manipulated using set operations, including union, intersection, and complementation, and satisfying the laws of an algebra of sets. Beyond that, operations like the converse of a relation and the composition of relations are available, satisfying the laws of a calculus of relations, for which there are textbooks by Ernst Schröder, Clarence Lewis, and Gunther Schmidt. A deeper analysis of relations involves decomposing them into subsets called , and placing them in a complete lattice. In some systems of axiomatic set theory, relations are extended to classes, which are generalizations of sets. This extension is needed for, among other things, modeling the concepts of "is an element of" or "is a subset of" in set theory, without running into logical inconsistencies such as Russell's paradox. The terms , dyadic relation and two-place relation are synonyms for binary relation, though some authors use the term "binary relation" for any subset of a Cartesian product without reference to and , and reserve the term "correspondence" for a binary relation with reference to and . Definition Given sets X and Y, the Cartesian product is defined as and its elements are called ordered pairs. A R over sets X and Y is a subset of The set X is called the or of R, and the set Y the or of R. In order to specify the choices of the sets X and Y, some authors define a or as an ordered triple , where G is a subset of called the of the binary relation. The statement reads "x is R-related to y" and is denoted by xRy. The or of R is the set of all x such that xRy for at least one y. The codomain of definition, , or of R is the set of all y such that xRy for at least one x. The of R is the union of its domain of definition and its codomain of definition. When a binary relation is called a (or ). To emphasize the fact that X and Y are allowed to be different, a binary relation is also called a heterogeneous relation. In a binary relation, the order of the elements is important; if then yRx can be true or false independently of xRy. For example, 3 divides 9, but 9 does not divide 3. Operations Union If R and S are binary relations over sets X and Y then is the of R and S over X and Y. The identity element is the empty relation. For example, is the union of < and =, and is the union of > and =. Intersection If R and S are binary relations over sets X and Y then is the of R and S over X and Y. The identity element is the universal relation. For example, the relation "is divisible by 6" is the intersection of the relations "is divisible by 3" and "is divisible by 2". Composition If R is a binary relation over sets X and Y, and S is a binary relation over sets Y and Z then (also denoted by ) is the of R and S over X and Z. The identity element is the identity relation. The order of R and S in the notation used here agrees with the standard notational order for composition of functions. For example, the composition (is parent of)(is mother of) yields (is maternal grandparent of), while the composition (is mother of)(is parent of) yields (is grandmother of). For the former case, if x is the parent of y and y is the mother of z, then x is the maternal grandparent of z. Converse If R is a binary relation over sets X and Y then is the of R over Y and X. For example, is the converse of itself, as is and and are each other's converse, as are and A binary relation is equal to its converse if and only if it is symmetric. Complement If R is a binary relation over sets X and Y then (also denoted by or ) is the of R over X and Y. For example, and are each other's complement, as are and and and and and, for total orders, also and and and The complement of the converse relation is the converse of the complement: If the complement has the following properties: If a relation is symmetric, then so is the complement. The complement of a reflexive relation is irreflexive—and vice versa. The complement of a strict weak order is a total preorder—and vice versa. Restriction If R is a binary homogeneous relation over a set X and S is a subset of X then is the of R to S over X. If R is a binary relation over sets X and Y and if S is a subset of X then is the of R to S over X and Y. If R is a binary relation over sets X and Y and if S is a subset of Y then is the of R to S over X and Y. If a relation is reflexive, irreflexive, symmetric, antisymmetric, asymmetric, transitive, total, trichotomous, a partial order, total order, strict weak order, total preorder (weak order), or an equivalence relation, then so too are its restrictions. However, the transitive closure of a restriction is a subset of the restriction of the transitive closure, i.e., in general not equal. For example, restricting the relation "x is parent of y" to females yields the relation "x is mother of the woman y"; its transitive closure does not relate a woman with her paternal grandmother. On the other hand, the transitive closure of "is parent of" is "is ancestor of"; its restriction to females does relate a woman with her paternal grandmother. Also, the various concepts of completeness (not to be confused with being "total") do not carry over to restrictions. For example, over the real numbers a property of the relation is that every non-empty subset with an upper bound in has a least upper bound (also called supremum) in However, for the rational numbers this supremum is not necessarily rational, so the same property does not hold on the restriction of the relation to the rational numbers. A binary relation R over sets X and Y is said to be a relation S over X and Y, written if R is a subset of S, that is, for all and if xRy, then xSy. If R is contained in S and S is contained in R, then R and S are called written R = S. If R is contained in S but S is not contained in R, then R is said to be than S, written For example, on the rational numbers, the relation is smaller than and equal to the composition Matrix representation Binary relations over sets X and Y can be represented algebraically by logical matrices indexed by X and Y with entries in the Boolean semiring (addition corresponds to OR and multiplication to AND) where matrix addition corresponds to union of relations, matrix multiplication corresponds to composition of relations (of a relation over X and Y and a relation over Y and Z), the Hadamard product corresponds to intersection of relations, the zero matrix corresponds to the empty relation, and the matrix of ones corresponds to the universal relation. Homogeneous relations (when ) form a matrix semiring (indeed, a matrix semialgebra over the Boolean semiring) where the identity matrix corresponds to the identity relation. Examples Special types of binary relations Some important types of binary relations R over sets X and Y are listed below. Uniqueness properties: Injective (also called left-unique): for all and all if and then . For such a relation, {Y} is called a primary key of R. For example, the green and blue binary relations in the diagram are injective, but the red one is not (as it relates both −1 and 1 to 1), nor the black one (as it relates both −1 and 1 to 0). Functional (also called right-unique, right-definite or univalent): for all and all if and then . Such a binary relation is called a . For such a relation, is called of R. For example, the red and green binary relations in the diagram are functional, but the blue one is not (as it relates 1 to both −1 and 1), nor the black one (as it relates 0 to both −1 and 1). One-to-one: injective and functional. For example, the green binary relation in the diagram is one-to-one, but the red, blue and black ones are not. One-to-many: injective and not functional. For example, the blue binary relation in the diagram is one-to-many, but the red, green and black ones are not. Many-to-one: functional and not injective. For example, the red binary relation in the diagram is many-to-one, but the green, blue and black ones are not. Many-to-many: not injective nor functional. For example, the black binary relation in the diagram is many-to-many, but the red, green and blue ones are not. Totality properties (only definable if the domain X and codomain Y are specified): Total (also called left-total): for all x in X there exists a y in Y such that . In other words, the domain of definition of R is equal to X. This property, is different from the definition of (also called by some authors) in Properties. Such a binary relation is called a . For example, the red and green binary relations in the diagram are total, but the blue one is not (as it does not relate −1 to any real number), nor the black one (as it does not relate 2 to any real number). As another example, > is a total relation over the integers. But it is not a total relation over the positive integers, because there is no in the positive integers such that . However, < is a total relation over the positive integers, the rational numbers and the real numbers. Every reflexive relation is total: for a given , choose . Surjective (also called right-total or onto): for all y in Y, there exists an x in X such that xRy. In other words, the codomain of definition of R is equal to Y. For example, the green and blue binary relations in the diagram are surjective, but the red one is not (as it does not relate any real number to −1), nor the black one (as it does not relate any real number to 2). Uniqueness and totality properties (only definable if the domain X and codomain Y are specified): A : a binary relation that is functional and total. For example, the red and green binary relations in the diagram are functions, but the blue and black ones are not. An : a function that is injective. For example, the green binary relation in the diagram is an injection, but the red, blue and black ones are not. A : a function that is surjective. For example, the green binary relation in the diagram is a surjection, but the red, blue and black ones are not. A : a function that is injective and surjective. For example, the green binary relation in the diagram is a bijection, but the red, blue and black ones are not. If relations over proper classes are allowed: Set-like (or ): for all in , the class of all in such that , i.e. , is a set. For example, the relation is set-like, and every relation on two sets is set-like. The usual ordering < over the class of ordinal numbers is a set-like relation, while its inverse > is not. Sets versus classes Certain mathematical "relations", such as "equal to", "subset of", and "member of", cannot be understood to be binary relations as defined above, because their domains and codomains cannot be taken to be sets in the usual systems of axiomatic set theory. For example, to model the general concept of "equality" as a binary relation take the domain and codomain to be the "class of all sets", which is not a set in the usual set theory. In most mathematical contexts, references to the relations of equality, membership and subset are harmless because they can be understood implicitly to be restricted to some set in the context. The usual work-around to this problem is to select a "large enough" set A, that contains all the objects of interest, and work with the restriction =A instead of =. Similarly, the "subset of" relation needs to be restricted to have domain and codomain P(A) (the power set of a specific set A): the resulting set relation can be denoted by Also, the "member of" relation needs to be restricted to have domain A and codomain P(A) to obtain a binary relation that is a set. Bertrand Russell has shown that assuming to be defined over all sets leads to a contradiction in naive set theory, see Russell's paradox. Another solution to this problem is to use a set theory with proper classes, such as NBG or Morse–Kelley set theory, and allow the domain and codomain (and so the graph) to be proper classes: in such a theory, equality, membership, and subset are binary relations without special comment. (A minor modification needs to be made to the concept of the ordered triple , as normally a proper class cannot be a member of an ordered tuple; or of course one can identify the binary relation with its graph in this context.) With this definition one can for instance define a binary relation over every set and its power set. Homogeneous relation A homogeneous relation over a set X is a binary relation over X and itself, i.e. it is a subset of the Cartesian product It is also simply called a (binary) relation over X. A homogeneous relation R over a set X may be identified with a directed simple graph permitting loops, where X is the vertex set and R is the edge set (there is an edge from a vertex x to a vertex y if and only if ). The set of all homogeneous relations over a set X is the power set which is a Boolean algebra augmented with the involution of mapping of a relation to its converse relation. Considering composition of relations as a binary operation on , it forms a semigroup with involution. Some important properties that a homogeneous relation over a set may have are: : for all . For example, is a reflexive relation but > is not. : for all not . For example, is an irreflexive relation, but is not. : for all if then . For example, "is a blood relative of" is a symmetric relation. : for all if and then For example, is an antisymmetric relation. : for all if then not . A relation is asymmetric if and only if it is both antisymmetric and irreflexive. For example, > is an asymmetric relation, but is not. : for all if and then . A transitive relation is irreflexive if and only if it is asymmetric. For example, "is ancestor of" is a transitive relation, while "is parent of" is not. : for all if then or . : for all or . : for all if then some exists such that and . A is a relation that is reflexive, antisymmetric, and transitive. A is a relation that is irreflexive, antisymmetric, and transitive. A is a relation that is reflexive, antisymmetric, transitive and connected. A is a relation that is irreflexive, antisymmetric, transitive and connected. An is a relation that is reflexive, symmetric, and transitive. For example, "x divides y" is a partial, but not a total order on natural numbers "x < y" is a strict total order on and "x is parallel to y" is an equivalence relation on the set of all lines in the Euclidean plane. All operations defined in section also apply to homogeneous relations. Beyond that, a homogeneous relation over a set X may be subjected to closure operations like: the smallest reflexive relation over X containing R, the smallest transitive relation over X containing R, the smallest equivalence relation over X containing R. Heterogeneous relation In mathematics, a heterogeneous relation is a binary relation, a subset of a Cartesian product where A and B are possibly distinct sets. The prefix hetero is from the Greek ἕτερος (heteros, "other, another, different"). A heterogeneous relation has been called a rectangular relation, suggesting that it does not have the square-symmetry of a homogeneous relation on a set where Commenting on the development of binary relations beyond homogeneous relations, researchers wrote, "...a variant of the theory has evolved that treats relations from the very beginning as or , i.e. as relations where the normal case is that they are relations between different sets." Calculus of relations Developments in algebraic logic have facilitated usage of binary relations. The calculus of relations includes the algebra of sets, extended by composition of relations and the use of converse relations. The inclusion meaning that aRb implies aSb, sets the scene in a lattice of relations. But since the inclusion symbol is superfluous. Nevertheless, composition of relations and manipulation of the operators according to Schröder rules, provides a calculus to work in the power set of In contrast to homogeneous relations, the composition of relations operation is only a partial function. The necessity of matching range to domain of composed relations has led to the suggestion that the study of heterogeneous relations is a chapter of category theory as in the category of sets, except that the morphisms of this category are relations. The of the category Rel are sets, and the relation-morphisms compose as required in a category. Induced concept lattice Binary relations have been described through their induced concept lattices: A concept C ⊂ R satisfies two properties: (1) The logical matrix of C is the outer product of logical vectors logical vectors. (2) C is maximal, not contained in any other outer product. Thus C is described as a non-enlargeable rectangle. For a given relation the set of concepts, enlarged by their joins and meets, forms an "induced lattice of concepts", with inclusion forming a preorder. The MacNeille completion theorem (1937) (that any partial order may be embedded in a complete lattice) is cited in a 2013 survey article "Decomposition of relations on concept lattices". The decomposition is where f and g are functions, called or left-total, univalent relations in this context. The "induced concept lattice is isomorphic to the cut completion of the partial order E that belongs to the minimal decomposition (f, g, E) of the relation R." Particular cases are considered below: E total order corresponds to Ferrers type, and E identity corresponds to difunctional, a generalization of equivalence relation on a set. Relations may be ranked by the Schein rank which counts the number of concepts necessary to cover a relation. Structural analysis of relations with concepts provides an approach for data mining. Particular relations Proposition: If R is a serial relation and RT is its transpose, then where is the m × m identity relation. Proposition: If R is a surjective relation, then where is the identity relation. Difunctional The idea of a difunctional relation is to partition objects by distinguishing attributes, as a generalization of the concept of an equivalence relation. One way this can be done is with an intervening set of indicators. The partitioning relation is a composition of relations using relations Jacques Riguet named these relations difunctional since the composition F GT involves univalent relations, commonly called partial functions. In 1950 Rigeut showed that such relations satisfy the inclusion: In automata theory, the term rectangular relation has also been used to denote a difunctional relation. This terminology recalls the fact that, when represented as a logical matrix, the columns and rows of a difunctional relation can be arranged as a block matrix with rectangular blocks of ones on the (asymmetric) main diagonal. More formally, a relation on is difunctional if and only if it can be written as the union of Cartesian products , where the are a partition of a subset of and the likewise a partition of a subset of . Using the notation {y: xRy} = xR, a difunctional relation can also be characterized as a relation R such that wherever x1R and x2R have a non-empty intersection, then these two sets coincide; formally implies In 1997 researchers found "utility of binary decomposition based on difunctional dependencies in database management." Furthermore, difunctional relations are fundamental in the study of bisimulations. In the context of homogeneous relations, a partial equivalence relation is difunctional. Ferrers type A strict order on a set is a homogeneous relation arising in order theory. In 1951 Jacques Riguet adopted the ordering of a partition of an integer, called a Ferrers diagram, to extend ordering to binary relations in general. The corresponding logical matrix of a general binary relation has rows which finish with a sequence of ones. Thus the dots of a Ferrer's diagram are changed to ones and aligned on the right in the matrix. An algebraic statement required for a Ferrers type relation R is If any one of the relations is of Ferrers type, then all of them are. Contact Suppose B is the power set of A, the set of all subsets of A. Then a relation g is a contact relation if it satisfies three properties: The set membership relation, ε = "is an element of", satisfies these properties so ε is a contact relation. The notion of a general contact relation was introduced by Georg Aumann in 1970. In terms of the calculus of relations, sufficient conditions for a contact relation include where is the converse of set membership (). Preorder R\R Every relation R generates a preorder which is the left residual. In terms of converse and complements, Forming the diagonal of , the corresponding row of and column of will be of opposite logical values, so the diagonal is all zeros. Then so that is a reflexive relation. To show transitivity, one requires that Recall that is the largest relation such that Then (repeat) (Schröder's rule) (complementation) (definition) The inclusion relation Ω on the power set of U can be obtained in this way from the membership relation on subsets of U: Fringe of a relation Given a relation R, a sub-relation called its is defined as When R is a partial identity relation, difunctional, or a block diagonal relation, then fringe(R) = R. Otherwise the fringe operator selects a boundary sub-relation described in terms of its logical matrix: fringe(R) is the side diagonal if R is an upper right triangular linear order or strict order. Fringe(R) is the block fringe if R is irreflexive () or upper right block triangular. Fringe(R) is a sequence of boundary rectangles when R is of Ferrers type. On the other hand, Fringe(R) = ∅ when R is a dense, linear, strict order. Mathematical heaps Given two sets A and B, the set of binary relations between them can be equipped with a ternary operation where bT denotes the converse relation of b. In 1953 Viktor Wagner used properties of this ternary operation to define semiheaps, heaps, and generalized heaps. The contrast of heterogeneous and homogeneous relations is highlighted by these definitions: See also Abstract rewriting system Additive relation, a many-valued homomorphism between modules Allegory (category theory) Category of relations, a category having sets as objects and binary relations as morphisms Confluence (term rewriting), discusses several unusual but fundamental properties of binary relations Correspondence (algebraic geometry), a binary relation defined by algebraic equations Hasse diagram, a graphic means to display an order relation Incidence structure, a heterogeneous relation between set of points and lines Logic of relatives, a theory of relations by Charles Sanders Peirce Order theory, investigates properties of order relations Notes References Bibliography Ernst Schröder (1895) Algebra der Logik, Band III, via Internet Archive External links
4092
https://en.wikipedia.org/wiki/Bishop
Bishop
A bishop is an ordained member of the clergy who is entrusted with a position of authority and oversight in a religious institution. In Christianity, bishops are normally responsible for the governance of dioceses. The role or office of bishop is called episcopacy. Organizationally, several Christian denominations utilize ecclesiastical structures that call for the position of bishops, while other denominations have dispensed with this office, seeing it as a symbol of power. Bishops have also exercised political authority. Traditionally, bishops claim apostolic succession, a direct historical lineage dating back to the original Twelve Apostles or Saint Paul. The bishops are by doctrine understood as those who possess the full priesthood given by Jesus Christ, and therefore may ordain other clergy, including other bishops. A person ordained as a deacon, priest (i.e. presbyter), and then bishop is understood to hold the fullness of the ministerial priesthood, given responsibility by Christ to govern, teach and sanctify the Body of Christ (the Church). Priests, deacons and lay ministers co-operate and assist their bishops in pastoral ministry. Some Pentecostal and other Protestant denominations have bishops who oversee congregations, though they do not claim apostolic succession. Terminology The English term bishop derives from the Greek word , meaning "overseer"; Greek was the language of the early Christian church. However, the term did not originate in Christianity. In Greek literature, the term had been used for several centuries before the advent of Christianity. It later transformed into the Latin , Old English , Middle English and lastly bishop. In the early Christian era the term was not always clearly distinguished from (literally: "elder" or "senior", origin of the modern English word priest), but is used in the sense of the order or office of bishop, distinct from that of presbyter, in the writings attributed to Ignatius of Antioch (died ). History in Christianity The earliest organization of the Church in Jerusalem was, according to most scholars, similar to that of Jewish synagogues, but it had a council or college of ordained presbyters (). In Acts 11:30 and Acts 15:22, a collegiate system of government in Jerusalem is chaired by James the Just, according to tradition the first bishop of the city. In Acts 14:23, the Apostle Paul ordains presbyters in churches in Anatolia. The word presbyter was not yet distinguished from overseer (, later used exclusively to mean bishop), as in Acts 20:17, Titus 1:5–7 and 1 Peter 5:1. The earliest writings of the Apostolic Fathers, the Didache and the First Epistle of Clement, for example, show the church used two terms for local church offices—presbyters (seen by many as an interchangeable term with or overseer) and deacon. In the First epistle to Timothy and Epistle to Titus in the New Testament a more clearly defined episcopate can be seen. Both letters state that Paul had left Timothy in Ephesus and Titus in Crete to oversee the local church. Paul commands Titus to ordain presbyters/bishops and to exercise general oversight. Early sources are unclear but various groups of Christian communities may have had the bishop surrounded by a group or college functioning as leaders of the local churches. Eventually the head or "monarchic" bishop came to rule more clearly, and all local churches would eventually follow the example of the other churches and structure themselves after the model of the others with the one bishop in clearer charge, though the role of the body of presbyters remained important. Eventually, as Christendom grew, bishops no longer directly served individual congregations. Instead, the metropolitan bishop (the bishop in a large city) appointed priests to minister each congregation, acting as the bishop's delegate. Apostolic Fathers Around the end of the 1st century, the church's organization became clearer in historical documents. In the works of the Apostolic Fathers, and Ignatius of Antioch in particular, the role of the episkopos, or bishop, became more important or, rather, already was very important and being clearly defined. While Ignatius of Antioch offers the earliest clear description of monarchial bishops (a single bishop over all house churches in a city) he is an advocate of monepiscopal structure rather than describing an accepted reality. To the bishops and house churches to which he writes, he offers strategies on how to pressure house churches who do not recognize the bishop into compliance. Other contemporary Christian writers do not describe monarchial bishops, either continuing to equate them with the presbyters or speaking of (bishops, plural) in a city. As the Church continued to expand, new churches in important cities gained their own bishop. Churches in the regions outside an important city were served by Chorbishop, an official rank of bishops. However, soon, presbyters and deacons were sent from bishop of a city church. Gradually priests replaced the chorbishops. Thus, in time, the bishop changed from being the leader of a single church confined to an urban area to being the leader of the churches of a given geographical area. Clement of Alexandria (end of the 2nd century) writes about the ordination of a certain Zachæus as bishop by the imposition of Simon Peter Bar-Jonah's hands. The words bishop and ordination are used in their technical meaning by the same Clement of Alexandria. The bishops in the 2nd century are defined also as the only clergy to whom the ordination to priesthood (presbyterate) and diaconate is entrusted: "a priest (presbyter) lays on hands, but does not ordain." (). At the beginning of the 3rd century, Hippolytus of Rome describes another feature of the ministry of a bishop, which is that of the : the primate of sacrificial priesthood and the power to forgive sins. Christian bishops and civil government The efficient organization of the Roman Empire became the template for the organisation of the church in the 4th century, particularly after Constantine's Edict of Milan. As the church moved from the shadows of privacy into the public forum it acquired land for churches, burials and clergy. In 391, Theodosius I decreed that any land that had been confiscated from the church by Roman authorities be returned. The most usual term for the geographic area of a bishop's authority and ministry, the diocese, began as part of the structure of the Roman Empire under Diocletian. As Roman authority began to fail in the western portion of the empire, the church took over much of the civil administration. This can be clearly seen in the ministry of two popes: Pope Leo I in the 5th century, and Pope Gregory I in the 6th century. Both of these men were statesmen and public administrators in addition to their role as Christian pastors, teachers and leaders. In the Eastern churches, latifundia entailed to a bishop's see were much less common, the state power did not collapse the way it did in the West, and thus the tendency of bishops acquiring civil power was much weaker than in the West. However, the role of Western bishops as civil authorities, often called prince bishops, continued throughout much of the Middle Ages. Bishops holding political office As well as being Archchancellors of the Holy Roman Empire after the 9th century, bishops generally served as chancellors to medieval monarchs, acting as head of the justiciary and chief chaplain. The Lord Chancellor of England was almost always a bishop up until the dismissal of Cardinal Thomas Wolsey by Henry VIII. Similarly, the position of Kanclerz in the Polish kingdom was always held by a bishop until the 16th century. In modern times, the principality of Andorra is headed by Co-Princes of Andorra, one of whom is the Bishop of Urgell and the other, the sitting President of France, an arrangement that began with the Paréage of Andorra (1278), and was ratified in the 1993 constitution of Andorra. The office of the Papacy is inherently held by the sitting Roman Catholic Bishop of Rome. Though not originally intended to hold temporal authority, since the Middle Ages the power of the Papacy gradually expanded deep into the secular realm and for centuries the sitting Bishop of Rome was the most powerful governmental office in Central Italy. In modern times, the Pope is also the sovereign Prince of Vatican City, an internationally recognized micro-state located entirely within the city of Rome. In France, prior to the Revolution, representatives of the clergy — in practice, bishops and abbots of the largest monasteries — comprised the First Estate of the Estates-General. This role was abolished after separation of Church and State was implemented during the French Revolution. In the 21st century, the more senior bishops of the Church of England continue to sit in the House of Lords of the Parliament of the United Kingdom, as representatives of the established church, and are known as Lords Spiritual. The Bishop of Sodor and Man, whose diocese lies outside the United Kingdom, is an ex officio member of the Legislative Council of the Isle of Man. In the past, the Bishop of Durham had extensive vice-regal powers within his northern diocese, which was a county palatine, the County Palatine of Durham, (previously, Liberty of Durham) of which he was ex officio the earl. In the 19th century, a gradual process of reform was enacted, with the majority of the bishop's historic powers vested in The Crown by 1858. Eastern Orthodox bishops, along with all other members of the clergy, are canonically forbidden to hold political office. Occasional exceptions to this rule are tolerated when the alternative is political chaos. In the Ottoman Empire, the Patriarch of Constantinople, for example, had de facto administrative, cultural and legal jurisdiction, as well as spiritual authority, over all Eastern Orthodox Christians of the empire, as part of the Ottoman millet system. An Orthodox bishop headed the Prince-Bishopric of Montenegro from 1516 to 1852, assisted by a secular guvernadur. More recently, Archbishop Makarios III of Cyprus, served as President of the Cyprus from 1960 to 1977, an extremely turbulent time period on the island. In 2001, Peter Hollingworth, AC, OBE – then the Anglican Archbishop of Brisbane – was controversially appointed Governor-General of Australia. Although Hollingworth gave up his episcopal position to accept the appointment, it still attracted considerable opposition in a country which maintains a formal separation between Church and State. Episcopacy during the English Civil War During the period of the English Civil War, the role of bishops as wielders of political power and as upholders of the established church became a matter of heated political controversy. Presbyterianism was the polity of most Reformed Churches in Europe, and had been favored by many in England since the English Reformation. Since in the primitive church the offices of presbyter and were not clearly distinguished, many Puritans held that this was the only form of government the church should have. The Anglican divine, Richard Hooker, objected to this claim in his famous work Of the Laws of Ecclesiastic Polity while, at the same time, defending Presbyterian ordination as valid (in particular Calvin's ordination of Beza). This was the official stance of the English Church until the Commonwealth, during which time, the views of Presbyterians and Independents (Congregationalists) were more freely expressed and practiced. Christian churches Catholic, Eastern Orthodox, Oriental Orthodox, Lutheran and Anglican churches Bishops form the leadership in the Catholic Church, the Eastern Orthodox Church, the Oriental Orthodox Churches, certain Lutheran Churches, the Anglican Communion, the Independent Catholic Churches, the Independent Anglican Churches, and certain other, smaller, denominations. The traditional role of a bishop is as pastor of a diocese (also called a bishopric, synod, eparchy or see), and so to serve as a "diocesan bishop", or "eparch" as it is called in many Eastern Christian churches. Dioceses vary considerably in size, geographically and population-wise. Some dioceses around the Mediterranean Sea which were Christianised early are rather compact, whereas dioceses in areas of rapid modern growth in Christian commitment—as in some parts of Sub-Saharan Africa, South America and the Far East—are much larger and more populous. As well as traditional diocesan bishops, many churches have a well-developed structure of church leadership that involves a number of layers of authority and responsibility. Duties In Catholicism, Eastern Orthodoxy, Oriental Orthodoxy, High Church Lutheranism, and Anglicanism, only a bishop can ordain other bishops, priests, and deacons. In the Eastern liturgical tradition, a priest can celebrate the Divine Liturgy only with the blessing of a bishop. In Byzantine usage, an antimension signed by the bishop is kept on the altar partly as a reminder of whose altar it is and under whose omophorion the priest at a local parish is serving. In Syriac Church usage, a consecrated wooden block called a thabilitho is kept for the same reasons. The bishop is the ordinary minister of the sacrament of confirmation in the Latin Church, and in the Old Catholic communion only a bishop may administer this sacrament. In the Lutheran and Anglican churches, the bishop normatively administers the rite of confirmation, although in those denominations that do not have an episcopal polity, confirmation is administered by the priest. However, in the Byzantine and other Eastern rites, whether Eastern or Oriental Orthodox or Eastern Catholic, chrismation is done immediately after baptism, and thus the priest is the one who confirms, using chrism blessed by a bishop. Ordination of Catholic, Eastern Orthodox, Oriental Orthodox, Lutheran and Anglican bishops Bishops in all of these communions are ordained by other bishops through the laying on of hands. Ordination of a bishop, and thus continuation of apostolic succession, takes place through a ritual centred on the imposition of hands and prayer. Catholic, Eastern Orthodox, Oriental Orthodox, Anglican, Old Catholic and some Lutheran bishops claim to be part of the continuous sequence of ordained bishops since the days of the apostles referred to as apostolic succession. In Scandinavia and the Baltic region, Lutheran churches participating in the Porvoo Communion (those of Iceland, Norway, Sweden, Finland, Estonia, and Lithuania), as well as many non-Porvoo membership Lutheran churches (including those of Kenya, Latvia, and Russia), as well as the confessional Communion of Nordic Lutheran Dioceses, believe that they ordain their bishops in the apostolic succession in lines stemming from the original apostles. The New Westminster Dictionary of Church History states that "In Sweden the apostolic succession was preserved because the Catholic bishops were allowed to stay in office, but they had to approve changes in the ceremonies." Peculiar to the Catholic Church While traditional teaching maintains that any bishop with apostolic succession can validly perform the ordination of another bishop, some churches require two or three bishops participate, either to ensure sacramental validity or to conform with church law. Catholic doctrine holds that one bishop can validly ordain another (priest) as a bishop. Though a minimum of three bishops participating is desirable (there are usually several more) in order to demonstrate collegiality, canonically only one bishop is necessary. The practice of only one bishop ordaining was normal in countries where the Church was persecuted under Communist rule. The title of archbishop or metropolitan may be granted to a senior bishop, usually one who is in charge of a large ecclesiastical jurisdiction. He may, or may not, have provincial oversight of suffragan bishops and may possibly have auxiliary bishops assisting him. Apart from the ordination, which is always done by other bishops, there are different methods as to the actual selection of a candidate for ordination as bishop. In the Catholic Church the Congregation for Bishops generally oversees the selection of new bishops with the approval of the pope. The papal nuncio usually solicits names from the bishops of a country, consults with priests and leading members of a laity, and then selects three to be forwarded to the Holy See. In Europe, some cathedral chapters have duties to elect bishops. The Eastern Catholic churches generally elect their own bishops. Most Eastern Orthodox churches allow varying amounts of formalised laity or lower clergy influence on the choice of bishops. This also applies in those Eastern churches which are in union with the pope, though it is required that he give assent. The pope, in addition to being the Bishop of Rome and spiritual head of the Catholic Church, is also the Patriarch of the Latin Church. Each bishop within the Latin Church is answerable directly to the Pope and not any other bishop except to metropolitans in certain oversight instances. The pope previously used the title Patriarch of the West, but this title was dropped from use in 2006, a move which caused some concern within the Eastern Orthodox Communion as, to them, it implied wider papal jurisdiction. Recognition of other churches' ordinations The Catholic Church does recognise as valid (though illicit) ordinations done by breakaway Catholic, Old Catholic or Oriental bishops, and groups descended from them; it also regards as both valid and licit those ordinations done by bishops of the Eastern churches, so long as those receiving the ordination conform to other canonical requirements (for example, is an adult male) and an eastern orthodox rite of episcopal ordination, expressing the proper functions and sacramental status of a bishop, is used; this has given rise to the phenomenon of (for example, clergy of the Independent Catholic groups which claim apostolic succession, though this claim is rejected by both Catholicism and Eastern Orthodoxy). With respect to Lutheranism, "the Catholic Church has never officially expressed its judgement on the validity of orders as they have been handed down by episcopal succession in these two national Lutheran churches" (the Evangelical Lutheran Church of Sweden and the Evangelical Lutheran Church of Finland) though it does "question how the ecclesiastical break in the 16th century has affected the apostolicity of the churches of the Reformation and thus the apostolicity of their ministry". Since Pope Leo XIII issued the bull in 1896, the Catholic Church has insisted that Anglican orders are invalid because of the Reformed changes in the Anglican ordination rites of the 16th century and divergence in understanding of the theology of priesthood, episcopacy and Eucharist. However, since the 1930s, Utrecht Old Catholic bishops (recognised by the Holy See as validly ordained) have sometimes taken part in the ordination of Anglican bishops. According to the writer Timothy Dufort, by 1969, all Church of England bishops had acquired Old Catholic lines of apostolic succession recognised by the Holy See. This development has been used to argue that the strain of apostolic succession has been re-introduced into Anglicanism, at least within the Church of England. However, other issues, such as the Anglican ordination of women, is at variance with Catholic understanding of Christian teaching, and have contributed to the reaffirmation of Catholic rejection of Anglican ordinations. The Eastern Orthodox Churches do not accept the validity of any ordinations performed by the Independent Catholic groups, as Eastern Orthodoxy considers to be spurious any consecration outside the Church as a whole. Eastern Orthodoxy considers apostolic succession to exist only within the Universal Church, and not through any authority held by individual bishops; thus, if a bishop ordains someone to serve outside the (Eastern Orthodox) Church, the ceremony is ineffectual, and no ordination has taken place regardless of the ritual used or the ordaining prelate's position within the Eastern Orthodox Churches. The position of the Catholic Church is slightly different. Whilst it does recognise the validity of the orders of certain groups which separated from communion with Holy See (for instance, the ordinations of the Old Catholics in communion with Utrecht, as well as the Polish National Catholic Church - which received its orders directly from Utrecht, and was until recently part of that communion), Catholicism does not recognise the orders of any group whose teaching is at variance with what they consider the core tenets of Christianity; this is the case even though the clergy of the Independent Catholic groups may use the proper ordination ritual. There are also other reasons why the Holy See does not recognise the validity of the orders of the Independent clergy: They hold that the continuing practice among many Independent clergy of one person receiving multiple ordinations in order to secure apostolic succession, betrays an incorrect and mechanistic theology of ordination. They hold that the practice within Independent groups of ordaining women (such as within certain member communities of the Anglican Communion) demonstrates an understanding of priesthood that they vindicate is totally unacceptable to the Catholic and Eastern Orthodox churches as they believe that the Universal Church does not possess such authority; thus, they uphold that any ceremonies performed by these women should be considered being sacramentally invalid. The theology of male clergy within the Independent movement is also suspect according to the Catholics, as they presumably approve of the ordination of females, and may have even undergone an (invalid) ordination ceremony conducted by a woman. Whilst members of the Independent Catholic movement take seriously the issue of valid orders, it is highly significant that the relevant Vatican Congregations tend not to respond to petitions from Independent Catholic bishops and clergy who seek to be received into communion with the Holy See, hoping to continue in some sacramental role. In those instances where the pope does grant reconciliation, those deemed to be clerics within the Independent Old Catholic movement are invariably admitted as laity and not priests or bishops. There is a mutual recognition of the validity of orders amongst Catholic, Eastern Orthodox, Old Catholic, Oriental Orthodox and Assyrian Church of the East churches. Some provinces of the Anglican Communion have begun ordaining women as bishops in recent decades – for example, England, Ireland, Scotland, Wales, the United States, Australia, New Zealand, Canada and Cuba. The first woman to be consecrated a bishop within Anglicanism was Barbara Harris, who was ordained in the United States in 1989. In 2006, Katharine Jefferts Schori, the Episcopal Bishop of Nevada, became the first woman to become the presiding bishop of the Episcopal Church. In the Evangelical Lutheran Church in America (ELCA) and the Evangelical Lutheran Church in Canada (ELCIC), the largest Lutheran Church bodies in the United States and Canada, respectively, and roughly based on the Nordic Lutheran national churches (similar to that of the Church of England), bishops are elected by Synod Assemblies, consisting of both lay members and clergy, for a term of six years, which can be renewed, depending upon the local synod's "constitution" (which is mirrored on either the ELCA or ELCIC's national constitution). Since the implementation of concordats between the ELCA and the Episcopal Church of the United States and the ELCIC and the Anglican Church of Canada, all bishops, including the presiding bishop (ELCA) or the national bishop (ELCIC), have been consecrated using the historic succession in line with bishops from the Evangelical Lutheran Church of Sweden, with at least one Anglican bishop serving as co-consecrator. Since going into ecumenical communion with their respective Anglican body, bishops in the ELCA or the ELCIC not only approve the "rostering" of all ordained pastors, diaconal ministers, and associates in ministry, but they serve as the principal celebrant of all pastoral ordination and installation ceremonies, diaconal consecration ceremonies, as well as serving as the "chief pastor" of the local synod, upholding the teachings of Martin Luther as well as the documentations of the Ninety-Five Theses and the Augsburg Confession. Unlike their counterparts in the United Methodist Church, ELCA and ELCIC synod bishops do not appoint pastors to local congregations (pastors, like their counterparts in the Episcopal Church, are called by local congregations). The presiding bishop of the ELCA and the national bishop of the ELCIC, the national bishops of their respective bodies, are elected for a single 6-year term and may be elected to an additional term. Although ELCA agreed with the Episcopal Church to limit ordination to the bishop "ordinarily", ELCA pastor-ordinators are given permission to perform the rites in "extraordinary" circumstance. In practice, "extraordinary" circumstance have included disagreeing with Episcopalian views of the episcopate, and as a result, ELCA pastors ordained by other pastors are not permitted to be deployed to Episcopal Churches (they can, however, serve in Presbyterian Church USA, United Methodist Church, Reformed Church in America, and Moravian Church congregations, as the ELCA is in full communion with these denominations). The Lutheran Church–Missouri Synod (LCMS) and the Wisconsin Evangelical Lutheran Synod (WELS), the second and third largest Lutheran bodies in the United States and the two largest Confessional Lutheran bodies in North America, do not follow an episcopal form of governance, settling instead on a form of quasi-congregationalism patterned off what they believe to be the practice of the early church. The second largest of the three predecessor bodies of the ELCA, the American Lutheran Church, was a congregationalist body, with national and synod presidents before they were re-titled as bishops (borrowing from the Lutheran churches in Germany) in the 1980s. With regard to ecclesial discipline and oversight, national and synod presidents typically function similarly to bishops in episcopal bodies. Methodism African Methodist Episcopal Church In the African Methodist Episcopal Church, "Bishops are the Chief Officers of the Connectional Organization. They are elected for life by a majority vote of the General Conference which meets every four years." Christian Methodist Episcopal Church In the Christian Methodist Episcopal Church in the United States, bishops are administrative superintendents of the church; they are elected by "delegate" votes for as many years deemed until the age of 74, then the bishop must retire. Among their duties, are responsibility for appointing clergy to serve local churches as pastor, for performing ordinations, and for safeguarding the doctrine and discipline of the Church. The General Conference, a meeting every four years, has an equal number of clergy and lay delegates. In each Annual Conference, CME bishops serve for four-year terms. CME Church bishops may be male or female. United Methodist Church In the United Methodist Church (the largest branch of Methodism in the world) bishops serve as administrative and pastoral superintendents of the church. They are elected for life from among the ordained elders (presbyters) by vote of the delegates in regional (called jurisdictional) conferences, and are consecrated by the other bishops present at the conference through the laying on of hands. In the United Methodist Church bishops remain members of the "Order of Elders" while being consecrated to the "Office of the Episcopacy". Within the United Methodist Church only bishops are empowered to consecrate bishops and ordain clergy. Among their most critical duties is the ordination and appointment of clergy to serve local churches as pastor, presiding at sessions of the Annual, Jurisdictional, and General Conferences, providing pastoral ministry for the clergy under their charge, and safeguarding the doctrine and discipline of the Church. Furthermore, individual bishops, or the Council of Bishops as a whole, often serve a prophetic role, making statements on important social issues and setting forth a vision for the denomination, though they have no legislative authority of their own. In all of these areas, bishops of the United Methodist Church function very much in the historic meaning of the term. According to the Book of Discipline of the United Methodist Church, a bishop's responsibilities are: In each Annual Conference, United Methodist bishops serve for four-year terms, and may serve up to three terms before either retirement or appointment to a new Conference. United Methodist bishops may be male or female, with Marjorie Matthews being the first woman to be consecrated a bishop in 1980. The collegial expression of episcopal leadership in the United Methodist Church is known as the Council of Bishops. The Council of Bishops speaks to the Church and through the Church into the world and gives leadership in the quest for Christian unity and interreligious relationships. The Conference of Methodist Bishops includes the United Methodist Council of Bishops plus bishops from affiliated autonomous Methodist or United Churches. John Wesley consecrated Thomas Coke a "General Superintendent", and directed that Francis Asbury also be consecrated for the United States of America in 1784, where the Methodist Episcopal Church first became a separate denomination apart from the Church of England. Coke soon returned to England, but Asbury was the primary builder of the new church. At first he did not call himself bishop, but eventually submitted to the usage by the denomination. Notable bishops in United Methodist history include Coke, Asbury, Richard Whatcoat, Philip William Otterbein, Martin Boehm, Jacob Albright, John Seybert, Matthew Simpson, John S. Stamm, William Ragsdale Cannon, Marjorie Matthews, Leontine T. Kelly, William B. Oden, Ntambo Nkulu Ntanda, Joseph Sprague, William Henry Willimon, and Thomas Bickerton. The Church of Jesus Christ of Latter-day Saints In the Church of Jesus Christ of Latter-day Saints, the Bishop is the leader of a local congregation, called a ward. As with most LDS priesthood holders, the bishop is a part-time lay minister and earns a living through other employment. As such, it is his duty to preside, call local leaders, and judge the worthiness of members for certain activities. The bishop does not deliver sermons at every service (generally asking members to do so), but is expected to be a spiritual guide for his congregation. It is therefore believed that he has both the right and ability to receive divine inspiration (through the Holy Spirit) for the ward under his direction. Because it is a part-time position, all able members are expected to assist in the management of the ward by holding delegated lay positions (for example, women's and youth leaders, teachers) referred to as callings. The bishop is especially responsible for leading the youth, in connection with the fact that a bishop is the president of the Aaronic priesthood in his ward (and is thus a form of Mormon Kohen). Although members are asked to confess serious sins to him, unlike the Catholic Church, he is not the instrument of divine forgiveness, but merely a guide through the repentance process (and a judge in case transgressions warrant excommunication or other official discipline). The bishop is also responsible for the physical welfare of the ward, and thus collects tithing and fast offerings and distributes financial assistance where needed. A literal descendant of Aaron has "legal right" to act as a bishop after being found worthy and ordained by the First Presidency. In the absence of a literal descendant of Aaron, a high priest in the Melchizedek priesthood is called to be a bishop. Each bishop is selected from resident members of the ward by the stake presidency with approval of the First Presidency, and chooses two counselors to form a bishopric. An priesthood holder called as bishop must be ordained a high priest if he is not already one, unlike the similar function of branch president. In special circumstances (such as a ward consisting entirely of young university students), a bishop may be chosen from outside the ward. Traditionally, bishops are married, though this is not always the case. A bishop is typically released after about five years and a new bishop is called to the position. Although the former bishop is released from his duties, he continues to hold the Aaronic priesthood office of bishop. Church members frequently refer to a former bishop as "Bishop" as a sign of respect and affection. Latter-day Saint bishops do not wear any special clothing or insignia the way clergy in many other churches do, but are expected to dress and groom themselves neatly and conservatively per their local culture, especially when performing official duties. Bishops (as well as other members of the priesthood) can trace their line of authority back to Joseph Smith, who, according to church doctrine, was ordained to lead the Church in modern times by the ancient apostles Peter, James, and John, who were ordained to lead the Church by Jesus Christ. At the global level, the presiding bishop oversees the temporal affairs (buildings, properties, commercial corporations, and so on) of the worldwide Church, including the Church's massive global humanitarian aid and social welfare programs. The presiding bishop has two counselors; the three together form the presiding bishopric. As opposed to ward bishoprics, where the counselors do not hold the office of bishop, all three men in the presiding bishopric hold the office of bishop, and thus the counselors, as with the presiding bishop, are formally referred to as "Bishop". Irvingism New Apostolic Church The New Apostolic Church (NAC) knows three classes of ministries: Deacons, Priests and Apostles. The Apostles, who are all included in the apostolate with the Chief Apostle as head, are the highest ministries. Of the several kinds of priest....ministries, the bishop is the highest. Nearly all bishops are set in line directly from the chief apostle. They support and help their superior apostle. Pentecostalism Church of God in Christ In the Church of God in Christ (COGIC), the ecclesiastical structure is composed of large dioceses that are called "jurisdictions" within COGIC, each under the authority of a bishop, sometimes called "state bishops". They can either be made up of large geographical regions of churches or churches that are grouped and organized together as their own separate jurisdictions because of similar affiliations, regardless of geographical location or dispersion. Each state in the U.S. has at least one jurisdiction while others may have several more, and each jurisdiction is usually composed of between 30 and 100 churches. Each jurisdiction is then broken down into several districts, which are smaller groups of churches (either grouped by geographical situation or by similar affiliations) which are each under the authority of District Superintendents who answer to the authority of their jurisdictional/state bishop. There are currently over 170 jurisdictions in the United States, and over 30 jurisdictions in other countries. The bishops of each jurisdiction, according to the COGIC Manual, are considered to be the modern day equivalent in the church of the early apostles and overseers of the New Testament church, and as the highest ranking clergymen in the COGIC, they are tasked with the responsibilities of being the head overseers of all religious, civil, and economic ministries and protocol for the church denomination. They also have the authority to appoint and ordain local pastors, elders, ministers, and reverends within the denomination. The bishops of the COGIC denomination are all collectively called "The Board of Bishops". From the Board of Bishops, and the General Assembly of the COGIC, the body of the church composed of clergy and lay delegates that are responsible for making and enforcing the bylaws of the denomination, every four years, twelve bishops from the COGIC are elected as "The General Board" of the church, who work alongside the delegates of the General Assembly and Board of Bishops to provide administration over the denomination as the church's head executive leaders. One of twelve bishops of the General Board is also elected the "presiding bishop" of the church, and two others are appointed by the presiding bishop himself, as his first and second assistant presiding bishops. Bishops in the Church of God in Christ usually wear black clergy suits which consist of a black suit blazer, black pants, a purple or scarlet clergy shirt and a white clerical collar, which is usually referred to as "Class B Civic attire". Bishops in COGIC also typically wear the Anglican Choir Dress style vestments of a long purple or scarlet chimere, cuffs, and tippet worn over a long white rochet, and a gold pectoral cross worn around the neck with the tippet. This is usually referred to as "Class A Ceremonial attire". The bishops of COGIC alternate between Class A Ceremonial attire and Class B Civic attire depending on the protocol of the religious services and other events they have to attend. Church of God (Cleveland, Tennessee) In the polity of the Church of God (Cleveland, Tennessee), the international leader is the presiding bishop, and the members of the executive committee are executive bishops. Collectively, they supervise and appoint national and state leaders across the world. Leaders of individual states and regions are administrative bishops, who have jurisdiction over local churches in their respective states and are vested with appointment authority for local pastorates. All ministers are credentialed at one of three levels of licensure, the most senior of which is the rank of ordained bishop. To be eligible to serve in state, national, or international positions of authority, a minister must hold the rank of ordained bishop. Pentecostal Church of God In 2002, the general convention of the Pentecostal Church of God came to a consensus to change the title of their overseer from general superintendent to bishop. The change was brought on because internationally, the term bishop is more commonly related to religious leaders than the previous title. The title bishop is used for both the general (international leader) and the district (state) leaders. The title is sometimes used in conjunction with the previous, thus becoming general (district) superintendent/bishop. Seventh-day Adventists According to the Seventh-day Adventist understanding of the doctrine of the Church: "The "elders" (Greek, ) or "bishops" () were the most important officers of the church. The term elder means older one, implying dignity and respect. His position was similar to that of the one who had supervision of the synagogue. The term bishop means "overseer". Paul used these terms interchangeably, equating elders with overseers or bishops (Acts 20:17,28; Titus 1:5, 7). "Those who held this position supervised the newly formed churches. Elder referred to the status or rank of the office, while bishop denoted the duty or responsibility of the office—"overseer". Since the apostles also called themselves elders (1 Peter 5:1; 2 John 1; 3 John 1), it is apparent that there were both local elders and itinerant elders, or elders at large. But both kinds of elder functioned as shepherds of the congregations." The above understanding is part of the basis of Adventist organizational structure. The world wide Seventh-day Adventist church is organized into local districts, conferences or missions, union conferences or union missions, divisions, and finally at the top is the general conference. At each level (with exception to the local districts), there is an elder who is elected president and a group of elders who serve on the executive committee with the elected president. Those who have been elected president would in effect be the "bishop" while never actually carrying the title or ordained as such because the term is usually associated with the episcopal style of church governance most often found in Catholic, Anglican, Methodist and some Pentecostal/Charismatic circles. Others Some Baptists also have begun taking on the title of bishop. In some smaller Protestant denominations and independent churches, the term bishop is used in the same way as pastor, to refer to the leader of the local congregation, and may be male or female. This usage is especially common in African-American churches in the US. In the Church of Scotland, which has a Presbyterian church structure, the word "bishop" refers to an ordained person, usually a normal parish minister, who has temporary oversight of a trainee minister. In the Presbyterian Church (USA), the term bishop is an expressive name for a Minister of Word and Sacrament who serves a congregation and exercises "the oversight of the flock of Christ." The term is traceable to the 1789 Form of Government of the PC (USA) and the Presbyterian understanding of the pastoral office. While not considered orthodox Christian, the Ecclesia Gnostica Catholica uses roles and titles derived from Christianity for its clerical hierarchy, including bishops who have much the same authority and responsibilities as in Catholicism. The Salvation Army does not have bishops but has appointed leaders of geographical areas, known as Divisional Commanders. Larger geographical areas, called Territories, are led by a Territorial Commander, who is the highest-ranking officer in that Territory. Jehovah's Witnesses do not use the title 'Bishop' within their organizational structure, but appoint elders to be overseers (to fulfill the role of oversight) within their congregations. The HKBP of Indonesia, the most prominent Protestant denomination in Indonesia, uses the term ephorus instead of bishop. In the Vietnamese syncretist religion of Caodaism, bishops () comprise the fifth of nine hierarchical levels, and are responsible for spiritual and temporal education as well as record-keeping and ceremonies in their parishes. At any one time there are seventy-two bishops. Their authority is described in Section I of the text (revealed through seances in December 1926). Caodai bishops wear robes and headgear of embroidered silk depicting the Divine Eye and the Eight Trigrams. (The color varies according to branch.) This is the full ceremonial dress; the simple version consists of a seven-layered turban. Dress and insignia in Christianity Traditionally, a number of items are associated with the office of a bishop, most notably the mitre, crosier, and ecclesiastical ring. Other vestments and insignia vary between Eastern and Western Christianity. In the Latin Church of the Catholic Church, the choir dress of a bishop includes the purple cassock with amaranth trim, rochet, purple zucchetto (skull cap), purple biretta, and pectoral cross. The cappa magna may be worn, but only within the bishop's own diocese and on especially solemn occasions. The mitre, zuchetto, and stole are generally worn by bishops when presiding over liturgical functions. For liturgical functions other than the Mass the bishop typically wears the cope. Within his own diocese and when celebrating solemnly elsewhere with the consent of the local ordinary, he also uses the crosier. When celebrating Mass, a bishop, like a priest, wears the chasuble. The Caeremoniale Episcoporum recommends, but does not impose, that in solemn celebrations a bishop should also wear a dalmatic, which can always be white, beneath the chasuble, especially when administering the sacrament of holy orders, blessing an abbot or abbess, and dedicating a church or an altar. The Caeremoniale Episcoporum no longer makes mention of episcopal gloves, episcopal sandals, liturgical stockings (also known as buskins), or the accoutrements that it once prescribed for the bishop's horse. The coat of arms of a Latin Church Catholic bishop usually displays a galero with a cross and crosier behind the escutcheon; the specifics differ by location and ecclesiastical rank (see Ecclesiastical heraldry). Anglican bishops generally make use of the mitre, crosier, ecclesiastical ring, purple cassock, purple zucchetto, and pectoral cross. However, the traditional choir dress of Anglican bishops retains its late mediaeval form, and looks quite different from that of their Catholic counterparts; it consists of a long rochet which is worn with a chimere. In the Eastern Churches (Eastern Orthodox, Eastern Rite Catholic) a bishop will wear the mandyas, panagia (and perhaps an enkolpion), sakkos, omophorion and an Eastern-style mitre. Eastern bishops do not normally wear an episcopal ring; the faithful kiss (or, alternatively, touch their forehead to) the bishop's hand. To seal official documents, he will usually use an inked stamp. An Eastern bishop's coat of arms will normally display an Eastern-style mitre, cross, eastern style crosier and a red and white (or red and gold) mantle. The arms of Oriental Orthodox bishops will display the episcopal insignia (mitre or turban) specific to their own liturgical traditions. Variations occur based upon jurisdiction and national customs. Cathedra In Catholic, Eastern Orthodox, Oriental Orthodox, Lutheran and Anglican cathedrals there is a special chair set aside for the exclusive use of the bishop. This is the bishop's cathedra and is often called the throne. In some Christian denominations, for example, the Anglican Communion, parish churches may maintain a chair for the use of the bishop when he visits; this is to signify the parish's union with the bishop. The term's use in non-Christian religions Buddhism The leader of the Buddhist Churches of America (BCA) is their bishop, which since April 23, 2020 is Marvin Harada. The Japanese title for the bishop of the BCA is , although the English title is favored over the Japanese. When it comes to many other Buddhist terms, the BCA chose to keep them in their original language (terms such as and ), but with some words (including ), they changed/translated these terms into English words. Between 1899 and 1944, the BCA held the name Buddhist Mission of North America. The leader of the Buddhist Mission of North America was called (superintendent/director) between 1899 and 1918. In 1918 the was promoted to bishop (). However, according to George J. Tanabe, the title "bishop" was in practice already used by Hawaiian Shin Buddhists (in Honpa Hongwanji Mission of Hawaii) even when the official title was kantoku. Bishops are also present in other Japanese Buddhist organizations. Higashi Hongan-ji's North American District, Honpa Honganji Mission of Hawaii, Jodo Shinshu Buddhist Temples of Canada, a Jodo Shu temple in Los Angeles, the Shingon temple Koyasan Buddhist Temple, Sōtō Mission in Hawai‘i (a Soto Zen Buddhist institution), and the Sōtō Zen Buddhist Community of South America () all have or have had leaders with the title bishop. As for the Sōtō Zen Buddhist Community of South America, the Japanese title is , but the leader is in practice referred to as "bishop". Tenrikyo Tenrikyo is a Japanese New Religion with influences from both Shinto and Buddhism. The leader of the Tenrikyo North American Mission has the title of bishop. See also Anglican ministry#Bishops Appointment of Catholic bishops Appointment of Church of England bishops Bishop in Europe Bishop in the Catholic Church Bishop of Alexandria, or Pope Bishops in the Church of Scotland Diocesan bishop Ecclesiastical polity (church governance) Congregationalist polity Presbyterian polity Ganzibra Gay bishops Hierarchy of the Catholic Church List of Catholic bishops of the United States List of Metropolitans and Patriarchs of Moscow List of types of spiritual teachers List of Lutheran bishops and archbishops Lists of patriarchs, archbishops, and bishops Lord Bishop Order of precedence in the Catholic Church Shepherd in religion Spokesperson bishops in the Church of England Suffragan Bishop in Europe Notes References Citations Sources External links Methodist/Anglican Thoughts On Apostolic Succession by Gregory Neal Methodist Episcopacy: In Search of Holy Orders by Gregory Neal The Old Catholic Church, Province of the United States The Ecumenical Catholic Communion* The United Methodist Church: Council of Bishops Vatican Website with Canon Law of Catholic Church Episcophobia: The Fear of bishops Christian terminology Ecclesiastical titles Episcopacy in Eastern Orthodoxy Episcopacy in Oriental Orthodoxy Anglican episcopal offices Methodism Religious leadership roles Bishop
4260
https://en.wikipedia.org/wiki/Dots%20and%20boxes
Dots and boxes
Dots and Boxes is a pencil-and-paper game for two players (sometimes more). It was first published in the 19th century by French mathematician Édouard Lucas, who called it . It has gone by many other names, including the dots and dashes, game of dots, dot to dot grid, boxes, and pigs in a pen. The game starts with an empty grid of dots. Usually two players take turns adding a single horizontal or vertical line between two unjoined adjacent dots. A player who completes the fourth side of a 1×1 box earns one point and takes another turn. A point is typically recorded by placing a mark that identifies the player in the box, such as an initial. The game ends when no more lines can be placed. The winner is the player with the most points. The board may be of any size grid. When short on time, or to learn the game, a 2×2 board (3×3 dots) is suitable. A 5×5 board, on the other hand, is good for experts. Strategy For most novice players, the game begins with a phase of more-or-less randomly connecting dots, where the only strategy is to avoid adding the third side to any box. This continues until all the remaining (potential) boxes are joined together into chains – groups of one or more adjacent boxes in which any move gives all the boxes in the chain to the opponent. At this point, players typically take all available boxes, then open the smallest available chain to their opponent. For example, a novice player faced with a situation like position 1 in the diagram on the right, in which some boxes can be captured, may take all the boxes in the chain, resulting in position 2. But with their last move, they have to open the next, larger chain, and the novice loses the game. A more experienced player faced with position 1 will instead play the double-cross strategy, taking all but 2 of the boxes in the chain and leaving position 3. The opponent will take these two boxes and then be forced to open the next chain. By achieving position 3, player A wins. The same double-cross strategy applies no matter how many long chains there are: a player using this strategy will take all but two boxes in each chain and take all the boxes in the last chain. If the chains are long enough, then this player will win. The next level of strategic complexity, between experts who would both use the double-cross strategy (if they were allowed to), is a battle for control: an expert player tries to force their opponent to open the first long chain, because the player who first opens a long chain usually loses. Against a player who does not understand the concept of a sacrifice, the expert simply has to make the correct number of sacrifices to encourage the opponent to hand them the first chain long enough to ensure a win. If the other player also sacrifices, the expert has to additionally manipulate the number of available sacrifices through earlier play. In combinatorial game theory, Dots and Boxes is an impartial game and many positions can be analyzed using Sprague–Grundy theory. However, Dots and Boxes lacks the normal play convention of most impartial games (where the last player to move wins), which complicates the analysis considerably. Unusual grids and variants Dots and Boxes need not be played on a rectangular gridit can be played on a triangular grid or a hexagonal grid. Dots and boxes has a dual graph form called "Strings-and-Coins". This game is played on a network of coins (vertices) joined by strings (edges). Players take turns cutting a string. When a cut leaves a coin with no strings, the player "pockets" the coin and takes another turn. The winner is the player who pockets the most coins. Strings-and-Coins can be played on an arbitrary graph. In analyses of Dots and Boxes, a game that starts with outer lines already drawn is called a Swedish board while the standard version that starts fully blank is called an American board. An intermediate version with only the left and bottom sides starting with drawn lines is called an Icelandic board. A related game is Dots, played by adding coloured dots to a blank grid, and joining them with straight or diagonal line in an attempt to surround an opponent's dots. References External links Abstract strategy games Mathematical games Paper-and-pencil games 1889 introductions
4403
https://en.wikipedia.org/wiki/BCS%20theory
BCS theory
BCS theory or Bardeen–Cooper–Schrieffer theory (named after John Bardeen, Leon Cooper, and John Robert Schrieffer) is the first microscopic theory of superconductivity since Heike Kamerlingh Onnes's 1911 discovery. The theory describes superconductivity as a microscopic effect caused by a condensation of Cooper pairs. The theory is also used in nuclear physics to describe the pairing interaction between nucleons in an atomic nucleus. It was proposed by Bardeen, Cooper, and Schrieffer in 1957; they received the Nobel Prize in Physics for this theory in 1972. History Rapid progress in the understanding of superconductivity gained momentum in the mid-1950s. It began with the 1948 paper, "On the Problem of the Molecular Theory of Superconductivity", where Fritz London proposed that the phenomenological London equations may be consequences of the coherence of a quantum state. In 1953, Brian Pippard, motivated by penetration experiments, proposed that this would modify the London equations via a new scale parameter called the coherence length. John Bardeen then argued in the 1955 paper, "Theory of the Meissner Effect in Superconductors", that such a modification naturally occurs in a theory with an energy gap. The key ingredient was Leon Cooper's calculation of the bound states of electrons subject to an attractive force in his 1956 paper, "Bound Electron Pairs in a Degenerate Fermi Gas". In 1957 Bardeen and Cooper assembled these ingredients and constructed such a theory, the BCS theory, with Robert Schrieffer. The theory was first published in April 1957 in the letter, "Microscopic theory of superconductivity". The demonstration that the phase transition is second order, that it reproduces the Meissner effect and the calculations of specific heats and penetration depths appeared in the December 1957 article, "Theory of superconductivity". They received the Nobel Prize in Physics in 1972 for this theory. In 1986, high-temperature superconductivity was discovered in La-Ba-Cu-O, at temperatures up to 30 K. Following experiments determined more materials with transition temperatures up to about 130 K, considerably above the previous limit of about 30 K. It is believed that BCS theory alone cannot explain this phenomenon and that other effects are in play. These effects are still not yet fully understood; it is possible that they even control superconductivity at low temperatures for some materials. Overview At sufficiently low temperatures, electrons near the Fermi surface become unstable against the formation of Cooper pairs. Cooper showed such binding will occur in the presence of an attractive potential, no matter how weak. In conventional superconductors, an attraction is generally attributed to an electron-lattice interaction. The BCS theory, however, requires only that the potential be attractive, regardless of its origin. In the BCS framework, superconductivity is a macroscopic effect which results from the condensation of Cooper pairs. These have some bosonic properties, and bosons, at sufficiently low temperature, can form a large Bose–Einstein condensate. Superconductivity was simultaneously explained by Nikolay Bogolyubov, by means of the Bogoliubov transformations. In many superconductors, the attractive interaction between electrons (necessary for pairing) is brought about indirectly by the interaction between the electrons and the vibrating crystal lattice (the phonons). Roughly speaking the picture is the following: An electron moving through a conductor will attract nearby positive charges in the lattice. This deformation of the lattice causes another electron, with opposite spin, to move into the region of higher positive charge density. The two electrons then become correlated. Because there are a lot of such electron pairs in a superconductor, these pairs overlap very strongly and form a highly collective condensate. In this "condensed" state, the breaking of one pair will change the energy of the entire condensate - not just a single electron, or a single pair. Thus, the energy required to break any single pair is related to the energy required to break all of the pairs (or more than just two electrons). Because the pairing increases this energy barrier, kicks from oscillating atoms in the conductor (which are small at sufficiently low temperatures) are not enough to affect the condensate as a whole, or any individual "member pair" within the condensate. Thus the electrons stay paired together and resist all kicks, and the electron flow as a whole (the current through the superconductor) will not experience resistance. Thus, the collective behavior of the condensate is a crucial ingredient necessary for superconductivity. Details BCS theory starts from the assumption that there is some attraction between electrons, which can overcome the Coulomb repulsion. In most materials (in low temperature superconductors), this attraction is brought about indirectly by the coupling of electrons to the crystal lattice (as explained above). However, the results of BCS theory do not depend on the origin of the attractive interaction. For instance, Cooper pairs have been observed in ultracold gases of fermions where a homogeneous magnetic field has been tuned to their Feshbach resonance. The original results of BCS (discussed below) described an s-wave superconducting state, which is the rule among low-temperature superconductors but is not realized in many unconventional superconductors such as the d-wave high-temperature superconductors. Extensions of BCS theory exist to describe these other cases, although they are insufficient to completely describe the observed features of high-temperature superconductivity. BCS is able to give an approximation for the quantum-mechanical many-body state of the system of (attractively interacting) electrons inside the metal. This state is now known as the BCS state. In the normal state of a metal, electrons move independently, whereas in the BCS state, they are bound into Cooper pairs by the attractive interaction. The BCS formalism is based on the reduced potential for the electrons' attraction. Within this potential, a variational ansatz for the wave function is proposed. This ansatz was later shown to be exact in the dense limit of pairs. Note that the continuous crossover between the dilute and dense regimes of attracting pairs of fermions is still an open problem, which now attracts a lot of attention within the field of ultracold gases. Underlying evidence The hyperphysics website pages at Georgia State University summarize some key background to BCS theory as follows: Evidence of a band gap at the Fermi level (described as "a key piece in the puzzle") the existence of a critical temperature and critical magnetic field implied a band gap, and suggested a phase transition, but single electrons are forbidden from condensing to the same energy level by the Pauli exclusion principle. The site comments that "a drastic change in conductivity demanded a drastic change in electron behavior". Conceivably, pairs of electrons might perhaps act like bosons instead, which are bound by different condensate rules and do not have the same limitation. Isotope effect on the critical temperature, suggesting lattice interactions The Debye frequency of phonons in a lattice is proportional to the inverse of the square root of the mass of lattice ions. It was shown that the superconducting transition temperature of mercury indeed showed the same dependence, by substituting natural mercury 202Hg with a different isotope 198Hg. An exponential rise in heat capacity near the critical temperature for some superconductors An exponential increase in heat capacity near the critical temperature also suggests an energy bandgap for the superconducting material. As superconducting vanadium is warmed toward its critical temperature, its heat capacity increases greatly in a very few degrees; this suggests an energy gap being bridged by thermal energy. The lessening of the measured energy gap towards the critical temperature This suggests a type of situation where some kind of binding energy exists but it is gradually weakened as the temperature increases toward the critical temperature. A binding energy suggests two or more particles or other entities that are bound together in the superconducting state. This helped to support the idea of bound particles - specifically electron pairs - and together with the above helped to paint a general picture of paired electrons and their lattice interactions. Implications BCS derived several important theoretical predictions that are independent of the details of the interaction, since the quantitative predictions mentioned below hold for any sufficiently weak attraction between the electrons and this last condition is fulfilled for many low temperature superconductors - the so-called weak-coupling case. These have been confirmed in numerous experiments: The electrons are bound into Cooper pairs, and these pairs are correlated due to the Pauli exclusion principle for the electrons, from which they are constructed. Therefore, in order to break a pair, one has to change energies of all other pairs. This means there is an energy gap for single-particle excitation, unlike in the normal metal (where the state of an electron can be changed by adding an arbitrarily small amount of energy). This energy gap is highest at low temperatures but vanishes at the transition temperature when superconductivity ceases to exist. The BCS theory gives an expression that shows how the gap grows with the strength of the attractive interaction and the (normal phase) single particle density of states at the Fermi level. Furthermore, it describes how the density of states is changed on entering the superconducting state, where there are no electronic states any more at the Fermi level. The energy gap is most directly observed in tunneling experiments and in reflection of microwaves from superconductors. BCS theory predicts the dependence of the value of the energy gap Δ at temperature T on the critical temperature Tc. The ratio between the value of the energy gap at zero temperature and the value of the superconducting transition temperature (expressed in energy units) takes the universal value independent of material. Near the critical temperature the relation asymptotes to which is of the form suggested the previous year by M. J. Buckingham based on the fact that the superconducting phase transition is second order, that the superconducting phase has a mass gap and on Blevins, Gordy and Fairbank's experimental results the previous year on the absorption of millimeter waves by superconducting tin. Due to the energy gap, the specific heat of the superconductor is suppressed strongly (exponentially) at low temperatures, there being no thermal excitations left. However, before reaching the transition temperature, the specific heat of the superconductor becomes even higher than that of the normal conductor (measured immediately above the transition) and the ratio of these two values is found to be universally given by 2.5. BCS theory correctly predicts the Meissner effect, i.e. the expulsion of a magnetic field from the superconductor and the variation of the penetration depth (the extent of the screening currents flowing below the metal's surface) with temperature. It also describes the variation of the critical magnetic field (above which the superconductor can no longer expel the field but becomes normal conducting) with temperature. BCS theory relates the value of the critical field at zero temperature to the value of the transition temperature and the density of states at the Fermi level. In its simplest form, BCS gives the superconducting transition temperature Tc in terms of the electron-phonon coupling potential V and the Debye cutoff energy ED: where N(0) is the electronic density of states at the Fermi level. For more details, see Cooper pairs. The BCS theory reproduces the isotope effect, which is the experimental observation that for a given superconducting material, the critical temperature is inversely proportional to the mass of the isotope used in the material. The isotope effect was reported by two groups on 24 March 1950, who discovered it independently working with different mercury isotopes, although a few days before publication they learned of each other's results at the ONR conference in Atlanta. The two groups are Emanuel Maxwell, and C. A. Reynolds, B. Serin, W. H. Wright, and L. B. Nesbitt. The choice of isotope ordinarily has little effect on the electrical properties of a material, but does affect the frequency of lattice vibrations. This effect suggests that superconductivity is related to vibrations of the lattice. This is incorporated into BCS theory, where lattice vibrations yield the binding energy of electrons in a Cooper pair. Little–Parks experiment - One of the first indications to the importance of the Cooper-pairing principle. See also Magnesium diboride, considered a BCS superconductor Quasiparticle Little–Parks effect, one of the first indications of the importance of the Cooper pairing principle. References Primary sources Further reading John Robert Schrieffer, Theory of Superconductivity, (1964), Michael Tinkham, Introduction to Superconductivity, Pierre-Gilles de Gennes, Superconductivity of Metals and Alloys, . Schmidt, Vadim Vasil'evich. The physics of superconductors: Introduction to fundamentals and applications. Springer Science & Business Media, 2013. External links ScienceDaily: Physicist Discovers Exotic Superconductivity (University of Arizona) August 17, 2006 Hyperphysics page on BCS BCS History Dance analogy of BCS theory as explained by Bob Schrieffer (audio recording) Mean-Field Theory: Hartree-Fock and BCS in E. Pavarini, E. Koch, J. van den Brink, and G. Sawatzky: Quantum materials: Experiments and Theory, Jülich 2016, Superconductivity
4544
https://en.wikipedia.org/wiki/Blind%20Willie%20McTell
Blind Willie McTell
Blind Willie McTell (born William Samuel McTier; May 5, 1898 – August 19, 1959) was a Piedmont blues and ragtime singer and guitarist. He played with a fluid, syncopated fingerstyle guitar technique, common among many exponents of Piedmont blues. Unlike his contemporaries, he came to use twelve-string guitars exclusively. McTell was also an adept slide guitarist, unusual among ragtime bluesmen. His vocal style, a smooth and often laid-back tenor, differed greatly from many of the harsher voices of Delta bluesmen such as Charley Patton. McTell performed in various musical styles, including blues, ragtime, religious music and hokum. McTell was born in Thomson, Georgia. He learned to play the guitar in his early teens. He soon became a street performer in several Georgia cities, including Atlanta and Augusta, and first recorded in 1927 for Victor Records. He never produced a major hit record, but he had a prolific recording career with different labels and under different names in the 1920s and 1930s. In 1940, he was recorded by the folklorist John A. Lomax and Ruby Terrill Lomax for the folk song archive of the Library of Congress. He was active in the 1940s and 1950s, playing on the streets of Atlanta, often with his longtime associate Curley Weaver. Twice more he recorded professionally. His last recordings originated during an impromptu session recorded by an Atlanta record store owner in 1956. McTell died three years later, having lived for years with diabetes and alcoholism. Despite his lack of commercial success, he was one of the few blues musicians of his generation who continued to actively play and record during the 1940s and 1950s. He did not live to see the American folk music revival, in which many other bluesmen were "rediscovered". McTell's influence extended over a wide variety of artists, including the Allman Brothers Band, who covered his "Statesboro Blues", and Bob Dylan, who paid tribute to him in his 1983 song "Blind Willie McTell", the refrain of which is "And I know no one can sing the blues like Blind Willie McTell". Other artists influenced by McTell include Taj Mahal, Alvin Youngblood Hart, Ralph McTell, Chris Smither, Jack White, and the White Stripes. Biography He was born William Samuel McTier in the Happy Valley community outside Thomson, Georgia. Most sources give the date of his birth as 1898, but researchers Bob Eagle and Eric LeBlanc suggest 1903, on the basis of his entry in the 1910 census. McTell was born blind in one eye and lost his remaining vision by late childhood. He attended schools for the blind in Georgia, New York and Michigan and showed proficiency in music from an early age, first playing the harmonica and accordion, learning to read and write music in Braille, and turning to the six-string guitar in his early teens. His family background was rich in music; both of his parents and an uncle played the guitar. He was related to the bluesman and gospel pioneer Thomas A. Dorsey. McTell's father left the family when Willie was young. After his mother died, in the 1920s, he left his hometown and became an itinerant musician, or "songster". He began his recording career in 1927 for Victor Records in Atlanta. McTell married Ruth Kate Williams, now better known as Kate McTell, in 1934. She accompanied him on stage and on several recordings before becoming a nurse in 1939. For most of their marriage, from 1942 until his death, they lived apart, she in Fort Gordon, near Augusta, and he working around Atlanta. In the years before World War II, McTell traveled and performed widely, recording for several labels under different names: Blind Willie McTell (for Victor and Decca), Blind Sammie (for Columbia), Georgia Bill (for Okeh), Hot Shot Willie (for Victor), Blind Willie (for Vocalion and Bluebird), Barrelhouse Sammie (for Atlantic), and Pig & Whistle Red (for Regal). The appellation "Pig & Whistle" was a reference to a chain of barbecue restaurants in Atlanta; McTell often played for tips in the parking lot of a Pig 'n Whistle restaurant. He also played behind a nearby building that later became Ray Lee's Blue Lantern Lounge. Like Lead Belly, another songster who began his career as a street artist, McTell favored the somewhat unwieldy and unusual twelve-string guitar, whose greater volume made it suitable for outdoor playing. In 1940 John A. Lomax and his wife, Ruby Terrill Lomax, a professor of classics at the University of Texas at Austin, interviewed and recorded McTell for the Archive of American Folk Song of the Library of Congress in a two-hour session held in their hotel room in Atlanta. These recordings document McTell's distinctive musical style, which bridges the gap between the raw country blues of the early part of the 20th century and the more conventionally melodious, ragtime-influenced East Coast Piedmont blues sound. The Lomaxes also elicited from the singer traditional songs (such as "The Boll Weevil" and "John Henry") and spirituals (such as "Amazing Grace"), which were not part of his usual commercial repertoire. In the interview, John A. Lomax is heard asking if McTell knows any "complaining" songs (an earlier term for protest songs), to which the singer replies somewhat uncomfortably and evasively that he does not. The Library of Congress paid McTell $10, the equivalent of $154.56 in 2011, for this two-hour session. The material from this 1940 session was issued in 1960 as an LP and later as a CD, under the somewhat misleading title The Complete Library of Congress Recordings, notwithstanding the fact that it was truncated, in that it omitted some of John A. Lomax's interactions with the singer and entirely omitted the contributions of Ruby Terrill Lomax. Ahmet Ertegun visited Atlanta in 1949 in search of blues artists for this new Atlantic Records label and after finding McTell playing on the street, arranged a recording session. Some of the songs were released on 78, but sold poorly. The complete session was released in 1972 as "Atlanta Twelve-String." McTell recorded for Regal Records in 1949, but these recordings also met with less commercial success than his previous works. He continued to perform around Atlanta, but his career was cut short by ill health, mostly due to diabetes and alcoholism. In 1956, an Atlanta record store manager, Edward Rhodes, discovered McTell playing in the street for quarters and enticed him with a bottle of corn liquor into his store, where he captured a few final performances on a tape recorder. These recordings were released posthumously by Prestige/Bluesville Records as Last Session. Beginning in 1957, McTell was a preacher at Mt. Zion Baptist Church in Atlanta. McTell died of a stroke in Milledgeville, Georgia, in 1959. He was buried at Jones Grove Church, near Thomson, Georgia, his birthplace. Author David Fulmer, who in 1992 was working on a documentary about McTell, paid to have a gravestone erected on his resting place. The name given on his gravestone is Willie Samuel McTier. He was inducted into the Blues Foundation's Blues Hall of Fame in 1981 and the Georgia Music Hall of Fame in 1990. In his recordings of "Lay Some Flowers on My Grave", "Lord, Send Me an Angel" and "Statesboro Blues", he pronounces his surname MacTell, with the stress on the first syllable. Influence McTell's most famous song, "Statesboro Blues", was first adapted by Taj Mahal with Jesse Ed Davis on slide guitar, then covered on an LP and frequently performed by the Allman Brothers Band; it also contributes to Canned Heat's "Goin' Up the Country". A short list of some of the artists who have performed the song includes Taj Mahal, David Bromberg, Dave Van Ronk, The Devil Makes Three and Ralph McTell, who changed his name on account of liking the song. Ry Cooder covered McTell's "Married Man's a Fool" on his 1973 album, Paradise and Lunch. Jack White, of the White Stripes considers McTell an influence; the White Stripes album De Stijl (2000) is dedicated to him and features a cover of his song "Southern Can Is Mine". The White Stripes also covered McTell's "Lord, Send Me an Angel", releasing it as a single in 2000. In 2013, Jack White's Third Man Records teamed up with Document Records to issue The Complete Recorded Works in Chronological Order of Charley Patton, Blind Willie McTell and the Mississippi Sheiks. Bob Dylan paid tribute to McTell on at least four occasions. In his 1965 song "Highway 61 Revisited", the second verse begins, "Georgia Sam he had a bloody nose", a reference to one of McTell's many recording names. (Note: There is no evidence of use of this moniker on any recordings) Dylan's song "Blind Willie McTell" was recorded in 1983 and released in 1991 on The Bootleg Series Volumes 1-3. Dylan also recorded covers of McTell's "Broke Down Engine" and "Delia" on his 1993 album, World Gone Wrong; Dylan's song "Po' Boy", on the album Love and Theft (2001), contains the lyric "had to go to Florida dodging them Georgia laws", which comes from McTell's "Kill It Kid". The Bath-based band Kill It Kid is named after the song of the same title. A billiards bar and concert in venue was named after McTell in the 1990s. The venue is now closed, but remains a fond memory for Georgia Southern University students at the time. Blind Willie's is a bar in the Virginia-Highlands neighborhood of Atlanta named after McTell that features blues musicians and bands. The Blind Willie McTell Blues Festival is held annually in Thomson, Georgia. Discography Singles As an accompanist Long-plays Selected compilations Blind Willie McTell 1927–1933: The Early Years, Yazoo L-1005 (1968) Blind Willie McTell 1949: Trying to Get Home, Biograph BLP-12008 (1969) King of the Georgia Blues Singers (1929–1935), Roots RL-324 (1969) Atlanta Twelve String, Atlantic SD-7224 (1972) Death Cell Blues, Biograph BLP-C-14 (1973) Blind Willie McTell: 1927–1935, Yazoo L-1037 (1974) Blind Willie McTell: 1927–1949, The Remaining Titles, Wolf WSE 102 (1982) Blues in the Dark, MCA 1368 (1983) Complete Recorded Works in Chronological Order, vol. 1, Document DOCD-5006 (1990) Complete Recorded Works in Chronological Order, vol. 2, Document DOCD-5007 (1990) Complete Recorded Works in Chronological Order, vol. 3, Document DOCD-5008 (1990) These three albums were issued together as the box set Statesboro Blues, Document DOCD-5677 (1990) Complete Library of Congress Recordings in Chronological Order, RST Blues Documents BDCD-6001 (1990) Pig 'n Whistle Red, Biograph BCD 126 (1993) The Definitive Blind Willie McTell, Legacy C2K-53234 (1994) The Classic Years 1927–1940, JSP JSP7711 (2003) King of the Georgia Blues, Snapper SBLUECD504X (2007) Selected compilations with other artists Blind Willie McTell/Memphis Minnie: Love Changin' Blues, Biograph BLP-12035 (1971) Atlanta Blues 1933, JEMF 106 (1979) Blind Willie McTell and Curley Weaver: The Post-War Years, RST Blues Documents BDCD 6014 (1990) Classic Blues Artwork from the 1920s, vol. 5, Blues Images – BIM-105 (2007) Footnotes Notes Citations Works cited General references Bastin, Bruce. Red River Blues: The Blues Tradition in the Southeast. Urbana and Chicago: University of Illinois Press, 1986, 1995. , . Charters, Samuel, ed. Sweet as the Showers of Rain. Oak Publications, 1977, pp, 120–131. External links New Georgia Encyclopedia – Blind Willie McTell article Illustrated Blind Willie McTell discography "Statesboro Blues" MP3 file on the Internet Archive David Fulmer, producer "Blind Willie's Blues" Documentary film, 1996 "The Dying Crapshooter's Blues" Novel by David Fulmer featuring McTell as a character John May interviews biographer Michael Gray Review of Hand Me My Travelin' Shoes: In Search of Blind Willie McTell by Michael Gray 1898 births 1959 deaths Year of birth uncertain African-American male singer-songwriters American acoustic guitarists American blues guitarists American blues harmonica players American blues singer-songwriters American male guitarists American street performers Blind musicians Blind singers Bluebird Records artists Columbia Records artists Country blues musicians East Coast blues musicians People from Thomson, Georgia Piedmont blues musicians Ragtime composers Songster musicians 20th-century American guitarists Guitarists from Georgia (U.S. state) Prestige Records artists Transatlantic Records artists Third Man Records artists African-American guitarists 20th-century African-American male singers Singer-songwriters from Georgia (U.S. state) American blind people
4692
https://en.wikipedia.org/wiki/Boston%20Corbett
Boston Corbett
Thomas H. "Boston" Corbett (January 29, 1832 – presumed dead September 1, 1894) was an American Union Army soldier who shot and killed U.S. president Abraham Lincoln's assassin, John Wilkes Booth. Corbett was initially arrested for disobeying orders, but was later released on the orders of Secretary of War Edwin Stanton, who referred to Corbett as "the patriot" upon dismissing him. He was largely considered a hero by the media and the public. Known for his devout religious beliefs and eccentric behavior, Corbett drifted around the United States before disappearing circa 1888. Circumstantial evidence suggests that he died in the Great Hinckley Fire in Minnesota in September 1894, although no period documentation has yet been found that undoubtedly identifies him as a victim of that devastating fire. Early life and education Corbett was born in London and immigrated with his family to New York City in 1840. The Corbetts moved frequently before eventually settling in Troy, New York. As a young man, Corbett began apprenticing as a milliner (also called a "hatter"), a profession that he would hold intermittently throughout his life. As a milliner, Corbett was regularly exposed to the fumes of mercury(II) nitrate, then used in the treatment of fur to produce felt used on hats. Excessive exposure to the compound can lead to hallucinations, psychosis and twitching (known as the "hatter's shakes"). Historians have theorized that the mental issues Corbett exhibited before and after the Civil War were caused by this exposure. Family and religion After working as a milliner in Troy, Corbett returned to New York City. He later married, but his wife and child died in childbirth. Following their deaths, he moved to Boston. Corbett became despondent over the loss of his wife and began drinking heavily. He was unable to hold a job and eventually became homeless. After a night of heavy drinking, he was confronted by a street preacher whose message persuaded him to join the Methodist Episcopal Church. Corbett immediately stopped drinking and became devoutly religious. After being baptized, he subsequently changed his name to Boston, the name of the city where he was converted. He regularly attended meetings at the Fulton and Bromfield Street churches where his enthusiastic behavior earned him the nickname "The Glory to God man". In an attempt to imitate Jesus, Corbett began to wear his hair very long (he was forced to cut it upon enlisting in the Union Army). In 1857, Corbett began working at a hat manufacturer's shop on Washington Street in downtown Boston. He was reported to be a proficient milliner, but was known to proselytize frequently and stop work to pray and sing for co-workers who used profanity in his presence. He also began working as a street preacher and would sermonize and distribute religious literature in North Square. Corbett soon earned a reputation around Boston for being a "local eccentric" and religious fanatic. On July 16, 1858, Corbett was propositioned by two prostitutes while walking home from a church meeting. He was deeply disturbed by the encounter. Upon returning to his room at a boardinghouse, Corbett began reading chapters 18 and 19 in the Gospel of Matthew ("And if thy right eye offend thee, pluck it out and cast it from thee....and there be eunuchs, which have made themselves eunuchs for the kingdom of heaven's sake"). In order to avoid sexual temptation and remain holy, he castrated himself with a pair of scissors. He then ate a meal and went to a prayer meeting before seeking medical treatment. Military career Enlistment in the Union army In April 1861, early in the American Civil War, Corbett enlisted as a private in Company I of the 12th Regiment New York Militia. Corbett's eccentric behavior quickly got him into trouble. He carried a Bible with him at all times and read passages aloud from it regularly, held unauthorized prayer meetings and argued with his superior officers. Corbett also condemned officers and superiors for what he perceived as violations of God's word. In one instance, he verbally reprimanded Colonel Daniel Butterfield for using profanity and taking the Lord's name in vain. He was sent to the guardhouse for several days but refused to apologize for his insubordination. Due to his continued disruptive behavior and refusal to take orders, Corbett was court-martialed and sentenced to be shot. His sentence was eventually reduced and he was discharged in August 1863. Corbett re-enlisted later that month as a private in Company L, 16th New York Cavalry Regiment. On June 24, 1864, he was captured by Confederate Colonel John S. Mosby's men in Culpeper, Virginia, and held prisoner at Andersonville prison for five months. While on the way to Andersonville, the following incident happened, told by a fellow prisoner of Corbett's named William Collins: Corbett was released in an exchange in November 1864 and was admitted to the Army hospital in Annapolis, Maryland, where he was treated for scurvy, malnutrition and exposure. On his return to his company, he was promoted to sergeant. Corbett later testified for the prosecution in the trial of the commandant of Andersonville, Captain Henry Wirz. Pursuit of John Wilkes Booth On April 24, 1865, Corbett's regiment was sent to apprehend John Wilkes Booth, the assassin of President Abraham Lincoln, whom Booth fatally shot on April 14, 1865. On April 26, the regiment surrounded Booth and one of his accomplices, David Herold, in a tobacco barn on the Virginia farm of Richard Garrett. Herold surrendered, but Booth refused and cried out, "I will not be taken alive!". The barn was set on fire in an attempt to force him out into the open, but Booth remained inside. Corbett was positioned near a large crack in the barn wall. Shooting Booth In an 1878 interview, Corbett claimed that he saw Booth aim his carbine, prompting him to shoot Booth with his Colt revolver despite Secretary of War Edwin Stanton's orders that Booth be captured alive. The bullet struck Booth in the back of the head behind his left ear, passed through his neck, and out into the barn. A low scream of pain like that produced by a sudden throttling came from the assassin, and he pitched headlong to the floor. Corbett and the other soldiers would note a sense of poetic, or cosmic, justice in that Lincoln and Booth were each shot around the same spot of the head. And the damage to Booth was no less severe than that to Lincoln: the bullet had pierced three vertebrae and partially severed his spinal cord, paralyzing him. Their conditions were different as well, as Mary Clemmer Ames summed it up, "The balls entered the skull of each at nearly the same spot, but the trifling difference made an immeasurable difference in the sufferings of the two. Mr. Lincoln was unconscious of all pain, while his assassin suffered as exquisite agony as if he had been broken on a wheel." Death of Booth In a weak voice, Booth asked for water and Lt. Colonel Everton Conger and Colonel Lafayette C. Baker gave it to him. A soldier poured water into his mouth, which he immediately spat out, unable to swallow. The bullet wound prevented him from swallowing any of the liquid. Booth asked them to roll him over and turn him facedown. Conger thought it a bad idea. "Then at least turn me on my side," the assassin pleaded. They did, but Conger saw that the move did not relieve Booth's suffering. Baker noticed it, too: "He seemed to suffer extreme pain whenever he was moved...and would several times repeat, 'Kill me.'" At sunrise, Booth remained in agonizing pain. His pulse weakened as his breathing became more labored and irregular. In agony, unable to move his limbs, he asked a soldier to lift his hands before his face and whispered as he gazed at them, "Useless ... Useless." These were his last words. A few minutes later, Booth began gasping for air as his throat continued to swell, and there was a shiver and a gurgle and his body shuddered, before he died from asphyxia. He died two hours after Corbett shot him. Conger initially thought Booth had shot himself. After realizing Booth had been shot by someone else, Conger and Lt. Doherty asked which officer had shot Booth. Corbett stepped forward and admitted he was the shooter. When asked why he had violated orders, Corbett replied, "Providence directed me". Court-martial He was immediately arrested and was accompanied by Lt. Doherty to the War Department in Washington, D.C. to be court-martialed. When questioned by Secretary Edwin Stanton about Booth's capture and shooting, both Doherty and Corbett himself agreed that Corbett had, in fact, disobeyed orders not to shoot. However, Corbett maintained that he believed Booth had intended to shoot his way out of the barn and that he acted in self-defense. He told Stanton, "...Booth would have killed me if I had not shot first. I think I did right." Corbett maintained that he didn't intend to kill Booth, but merely wanted to inflict a disabling wound, but either his aim slipped or Booth moved at the moment Corbett pulled the trigger. Stanton paused and then stated, "The rebel is dead. The patriot lives; he has spared the country expense, continued excitement and trouble. Discharge the patriot." Upon leaving the War Department, Corbett was greeted by a cheering crowd. As he made his way to Mathew Brady's studio to have his official portrait taken, the crowd followed him asking for autographs and requesting that he tell them about shooting Booth. Corbett told the crowd: Contradictions Eyewitnesses to Booth's shooting contradicted Corbett's version of events and expressed doubts that Corbett was responsible for shooting Booth. Officers who were near Corbett at the time claimed that they never saw him fire his gun (Corbett's gun was never inspected and was eventually lost). They claimed that Corbett came forward only after Lt. Colonel Conger asked who had shot Booth. Richard Garrett, the owner of the farm on which Booth was found, and his 12-year-old son Robert also contradicted Corbett's testimony that he acted in self-defense. Both maintained that Booth had never reached for his gun. While there was some criticism of Corbett's actions, he was largely considered a hero by the public and press. One newspaper editor declared that Corbett would, "live as one of the World's great avengers." For his part in Booth's capture, Corbett received a portion of the $100,000 reward money, amounting to $1,653.84 (). His annual salary as a U.S. sergeant was $204 (). Corbett received offers to purchase the gun he used to shoot Booth. He refused stating, "That is not mine—it belongs to the Government, and I would not sell it for any price." Corbett also declined an offer for one of Booth's pistols as he did not want a reminder of shooting Booth. Post-war life After his discharge from the army in August 1865, Corbett went back to work as a hatter in Boston and frequently attended the Bromfield Street Church. When the hatting business in Boston slowed, Corbett moved to Danbury, Connecticut, to continue his work and also "preached in the country round about." By 1870, he had relocated once again to Camden, New Jersey, where he was known as a "Methodist lay preacher". Corbett's inability to hold a job was attributed to his fanatical behavior; he was routinely fired after continuing his habit of stopping work to pray for his co-workers. In an effort to earn money, Corbett capitalized on his role as "Lincoln's Avenger". He gave lectures about the shooting of Booth accompanied by illustrated lantern slides at Sunday schools, women's groups and tent meetings. Corbett was never asked back due to his increasingly erratic behavior and incoherent speeches. R.B. Hoover, a man who later befriended Corbett, recalled that Corbett believed "men who were high in authority at Washington at the time of the assassination" were hounding him. Corbett said the men were angry because he had deprived them of prosecuting and executing John Wilkes Booth themselves. He also believed the same men had gotten him fired from various jobs. Corbett's paranoia was furthered by hate mail he received for killing Booth. He became fearful that "Booth's Avengers" or organizations like the "Secret Order" were planning to seek revenge upon him and took to carrying a pistol with him at all times. As his paranoia increased, Corbett began brandishing his pistol at friends or strangers he deemed suspicious. While attending the Soldiers' Reunion of the Blue and Gray in Caldwell, Ohio, in 1875, Corbett got into an argument with several men over the death of John Wilkes Booth. The men questioned if Booth had really been killed at all which enraged Corbett. He then drew his pistol on the men but was removed from the reunion before he could fire it. In 1878, Corbett moved to Concordia, Kansas, where he acquired a plot of land through homesteading upon which he constructed a dugout home. He continued working as a preacher and attended revival meetings frequently. In January 1887, Corbett was elected Assistant Doorkeeper in the Kansas House of Representatives where he frequently had run-ins with the public and elected officials, often brandishing his revolver. Presumed fate Due to his fame as "Lincoln's Avenger", Corbett was appointed assistant doorkeeper of the Kansas House of Representatives in Topeka in January 1887. On February 15, he became convinced that officers of the House were discriminating against him. He jumped to his feet, brandished a revolver and began chasing the officers out of the building. No one was hurt and Corbett was arrested. The following day, a judge declared Corbett insane and sent him to the Topeka Asylum for the Insane. On May 26, 1888, he escaped from the asylum on horseback. He then rode to Neodesha, Kansas, where he briefly stayed with Richard Thatcher, a man he had met while they were prisoners of war. When Corbett left, he told Thatcher he was going to Mexico. Rather than going to Mexico, Corbett is believed to have settled in a cabin he built in the forests near Hinckley, in Pine County in eastern Minnesota. He is believed to have died in the Great Hinckley Fire on September 1, 1894. Although there is no proof, the name "Thomas Corbett" appears on the list of dead and missing. Legacy Imposters In the years following Corbett's presumed death, several men came forward claiming to be "Lincoln's Avenger". A few years after Corbett was last seen in Neodesha, Kansas, a patent medicine salesman in Enid, Oklahoma, filed an application using Corbett's name to receive pension benefits. After an investigation proved that the man was not Boston Corbett, he was sent to prison. In September 1905, a man arrested in Dallas also claimed to be Corbett. He too was proven to be an imposter and was sent to prison for perjury, and then to the Government Hospital for the Insane. Memorials In 1958, Boy Scout Troop 31, of Concordia, Kansas, built a roadside monument to Corbett located on Key Road. A small sign was also placed to mark the dug hole where Corbett had lived for a time. See also List of people who disappeared Jack Ruby Notes References External links Boston Corbett: The Man Who Killed John Wilkes Booth Photo on Kansas Memory website 1832 births 1880s missing person cases American Civil War prisoners of war American escapees Methodists from Massachusetts Castrated people Converts to Methodism English emigrants to the United States American milliners Members of the Methodist Episcopal Church Methodist evangelists Military personnel from Troy, New York Missing person cases in Minnesota People associated with the assassination of Abraham Lincoln People declared dead in absentia People from Boston People from Camden, New Jersey People from Concordia, Kansas People from Hinckley, Minnesota Military personnel from London People from Noble County, Ohio People of New York (state) in the American Civil War Union Army soldiers Year of death unknown
4854
https://en.wikipedia.org/wiki/Bermuda%20Triangle
Bermuda Triangle
The Bermuda Triangle, also known as the Devil's Triangle, is an urban legend focused on a loosely defined region in the western part of the North Atlantic Ocean where a number of aircraft and ships are said to have disappeared under mysterious circumstances. The idea of the area as uniquely prone to disappearances arose in the mid-20th century, but most reputable sources dismiss the idea that there is any mystery. Origins The earliest suggestion of unusual disappearances in the Bermuda area appeared in a September 17, 1950, article published in The Miami Herald (Associated Press) by Edward Van Winkle Jones. Two years later, Fate magazine published "Sea Mystery at Our Back Door", a short article by George Sand covering the loss of several planes and ships, including the loss of Flight 19, a group of five US Navy Grumman TBM Avenger torpedo bombers on a training mission. Sand's article was the first to lay out the now-familiar triangular area where the losses took place, as well as the first to suggest a supernatural element to the Flight 19 incident. Flight 19 alone would be covered again in the April 1962 issue of American Legion magazine. In it, author Allan W. Eckert wrote that the flight leader had been heard saying, "We are entering white water, nothing seems right. We don't know where we are, the water is green, no white." He also wrote that officials at the Navy board of inquiry stated that the planes "flew off to Mars." In February 1964, Vincent Gaddis wrote an article called "The Deadly Bermuda Triangle" in the pulp magazine Argosy saying Flight 19 and other disappearances were part of a pattern of strange events in the region. The next year, Gaddis expanded this article into a book, Invisible Horizons. Other writers elaborated on Gaddis' ideas: John Wallace Spencer (Limbo of the Lost, 1969, repr. 1973); Charles Berlitz (The Bermuda Triangle, 1974); Richard Winer (The Devil's Triangle, 1974), and many others, all keeping to some of the same supernatural elements outlined by Eckert. Triangle area The Gaddis Argosy article delineated the boundaries of the triangle, giving its vertices as Miami; San Juan, Puerto Rico; and Bermuda. Subsequent writers did not necessarily follow this definition. Some writers gave different boundaries and vertices to the triangle, with the total area varying from . "Indeed, some writers even stretch it as far as the Irish coast." Consequently, the determination of which accidents occurred inside the triangle depends on which writer reported them. Criticism of the concept Larry Kusche Larry Kusche, author of The Bermuda Triangle Mystery: Solved (1975), argued that many claims of Gaddis and subsequent writers were exaggerated, dubious or unverifiable. Kusche's research revealed a number of inaccuracies and inconsistencies between Berlitz's accounts and statements from eyewitnesses, participants, and others involved in the initial incidents. Kusche noted cases where pertinent information went unreported, such as the disappearance of round-the-world yachtsman Donald Crowhurst, which Berlitz had presented as a mystery, despite clear evidence to the contrary. Another example was the ore-carrier recounted by Berlitz as lost without trace three days out of an Atlantic port when it had been lost three days out of a port with the same name in the Pacific Ocean. Kusche also argued that a large percentage of the incidents that sparked allegations of the Triangle's mysterious influence actually occurred well outside it. Often his research was simple: he would review period newspapers of the dates of reported incidents and find reports on possibly relevant events like unusual weather, that were never mentioned in the disappearance stories. Kusche concluded that: The number of ships and aircraft reported missing in the area was not significantly greater, proportionally speaking, than in any other part of the ocean. In an area frequented by tropical cyclones, the number of disappearances that did occur were, for the most part, neither disproportionate, unlikely, nor mysterious. Furthermore, Berlitz and other writers would often fail to mention such storms or even represent the disappearance as having happened in calm conditions when meteorological records clearly contradict this. The numbers themselves had been exaggerated by sloppy research. A boat's disappearance, for example, would be reported, but its eventual (if belated) return to port may not have been. Some disappearances had, in fact, never happened. One plane crash was said to have taken place in 1937, off Daytona Beach, Florida, in front of hundreds of witnesses. The legend of the Bermuda Triangle is a manufactured mystery, perpetuated by writers who either purposely or unknowingly made use of misconceptions, faulty reasoning, and sensationalism. In a 2013 study, the World Wide Fund for Nature identified the world's 10 most dangerous waters for shipping, but the Bermuda Triangle was not among them. Further responses When the UK Channel 4 television program The Bermuda Triangle (1992) was being produced by John Simmons of Geofilms for the Equinox series, the marine insurance market Lloyd's of London was asked if an unusually large number of ships had sunk in the Bermuda Triangle area. Lloyd's determined that large numbers of ships had not sunk there. Lloyd's does not charge higher rates for passing through this area. United States Coast Guard records confirm their conclusion. In fact, the number of supposed disappearances is relatively insignificant considering the number of ships and aircraft that pass through on a regular basis. The Coast Guard is also officially skeptical of the Triangle, noting that they collect and publish, through their inquiries, much documentation contradicting many of the incidents written about by the Triangle authors. In one such incident involving the 1972 explosion and sinking of the tanker , the Coast Guard photographed the wreck and recovered several bodies, in contrast with one Triangle author's claim that all the bodies had vanished, with the exception of the captain, who was found sitting in his cabin at his desk, clutching a coffee cup. In addition, V. A. Fogg sank off the coast of Texas, nowhere near the commonly accepted boundaries of the Triangle. The Nova/Horizon episode The Case of the Bermuda Triangle, aired on June 27, 1976, was highly critical, stating that "When we've gone back to the original sources or the people involved, the mystery evaporates. Science does not have to answer questions about the Triangle because those questions are not valid in the first place ... Ships and planes behave in the Triangle the same way they behave everywhere else in the world." Skeptical researchers, such as Ernest Taves and Barry Singer, have noted how mysteries and the paranormal are very popular and profitable. This has led to the production of vast amounts of material on topics such as the Bermuda Triangle. They were able to show that some of the pro-paranormal material is often misleading or inaccurate, but its producers continue to market it. Accordingly, they have claimed that the market is biased in favor of books, TV specials, and other media that support the Triangle mystery, and against well-researched material if it espouses a skeptical viewpoint. Benjamin Radford, an author and scientific paranormal investigator, noted in an interview on the Bermuda Triangle that it could be very difficult locating an aircraft lost at sea due to the vast search area, and although the disappearance might be mysterious, that did not make it paranormal or unexplainable. Radford further noted the importance of double-checking information as the mystery surrounding the Bermuda Triangle had been created by people who had neglected to do so. Hypothetical explanation attempts Persons accepting the Bermuda Triangle as a real phenomenon have offered a number of explanatory approaches. Paranormal explanations Triangle writers have used a number of supernatural concepts to explain the events. One explanation pins the blame on leftover technology from the mythical lost continent of Atlantis. Sometimes connected to the Atlantis story is the submerged rock formation known as the Bimini Road off the island of Bimini in the Bahamas, which is in the Triangle by some definitions. Followers of the purported psychic Edgar Cayce take his prediction that evidence of Atlantis would be found in 1968, as referring to the discovery of the Bimini Road. Believers describe the formation as a road, wall, or other structure, but the Bimini Road is of natural origin. Some hypothesize that a parallel universe exists in the Bermuda Triangle region, causing a time/space warp that sucks the objects around it into a parallel universe. Others attribute the events to UFOs. Charles Berlitz, author of various books on anomalous phenomena, lists several theories attributing the losses in the Triangle to anomalous or unexplained forces. Natural explanations Compass variations Compass problems are one of the cited phrases in many Triangle incidents. While some have theorized that unusual local magnetic anomalies may exist in the area, such anomalies have not been found. Compasses have natural magnetic variations in relation to the magnetic poles, a fact which navigators have known for centuries. Magnetic (compass) north and geographic (true) north are exactly the same only for a small number of places – for example, , in the United States, only those places on a line running from Wisconsin to the Gulf of Mexico. But the public may not be as informed, and think there is something mysterious about a compass "changing" across an area as large as the Triangle, which it naturally will. Gulf Stream The Gulf Stream (Florida Current) is a major surface current, primarily driven by thermohaline circulation that originates in the Gulf of Mexico and then flows through the Straits of Florida into the North Atlantic. In essence, it is a river within an ocean, and, like a river, it can and does carry floating objects. It has a maximum surface velocity of about . A small plane making a water landing or a boat having engine trouble can be carried away from its reported position by the current. Human error One of the most cited explanations in official inquiries as to the loss of any aircraft or vessel is human error. Human stubbornness may have caused businessman Harvey Conover to lose his sailing yacht, Revonoc, as he sailed into the teeth of a storm south of Florida on January 1, 1958. Violent weather Hurricanes are powerful storms that form in tropical waters and have historically cost thousands of lives and caused billions of dollars in damage. The sinking of Francisco de Bobadilla's Spanish fleet in 1502 was the first recorded instance of a destructive hurricane. These storms have in the past caused a number of incidents related to the Triangle. Many Atlantic hurricanes pass through the Triangle as they recurve off the Eastern Seaboard, and, before the advent of weather satellite, ships often had little to no warning of a hurricane's approach. A powerful downdraft of cold air was suspected to be a cause in the sinking of Pride of Baltimore on May 14, 1986. The crew of the sunken vessel noted the wind suddenly shifted and increased velocity from to . A National Hurricane Center satellite specialist, James Lushine, stated "during very unstable weather conditions the downburst of cold air from aloft can hit the surface like a bomb, exploding outward like a giant squall line of wind and water." A similar event occurred to Concordia in 2010, off the coast of Brazil. Methane hydrates An explanation for some of the disappearances has focused on the presence of large fields of methane hydrates (a form of natural gas) on the continental shelves. Laboratory experiments carried out in Australia have proven that bubbles can, indeed, sink a scale model ship by decreasing the density of the water; any wreckage consequently rising to the surface would be rapidly dispersed by the Gulf Stream. It has been hypothesized that periodic methane eruptions (sometimes called "mud volcanoes") may produce regions of frothy water that are no longer capable of providing adequate buoyancy for ships. If this were the case, such an area forming around a ship could cause it to sink very rapidly and without warning. Publications by the USGS describe large stores of undersea hydrates worldwide, including the Blake Ridge area, off the coast of the southeastern United States. However, according to the USGS, no large releases of gas hydrates are believed to have occurred in the Bermuda Triangle for the past 15,000 years. Notable incidents HMS Atalanta The sail training ship HMS Atalanta (originally named HMS Juno) disappeared with her entire crew after setting sail from the Royal Naval Dockyard, Bermuda for Falmouth, England on 31 January 1880. It was presumed that she sank in a powerful storm which crossed her route a couple of weeks after she sailed, and that her crew being composed primarily of inexperienced trainees may have been a contributing factor. The search for evidence of her fate attracted worldwide attention at the time (connection is also often made to the 1878 loss of the training ship HMS Eurydice, which foundered after departing the Royal Naval Dockyard in Bermuda for Portsmouth on 6 March), and she was alleged decades later to have been a victim of the mysterious triangle, an allegation resoundingly refuted by the research of author David Francis Raine in 1997. USS Cyclops The incident resulting in the single largest loss of life in the history of the US Navy not related to combat occurred when the collier Cyclops, carrying a full load of manganese ore and with one engine out of action, went missing without a trace with a crew of 309 sometime after March 4, 1918, after departing the island of Barbados. Although there is no strong evidence for any single theory, many independent theories exist, some blaming storms, some capsizing, and some suggesting that wartime enemy activity was to blame for the loss. In addition, two of Cyclopss sister ships, and , were subsequently lost in the North Atlantic during World War II. Both ships were transporting heavy loads of metallic ore similar to that which was loaded on Cyclops during her fatal voyage. In all three cases structural failure due to overloading with a much denser cargo than designed is considered the most likely cause of sinking. Carroll A. Deering Carroll A. Deering, a five-masted schooner built in 1919, was found hard aground and abandoned at Diamond Shoals, near Cape Hatteras, North Carolina, on January 31, 1921. FBI investigation into the Deering scrutinized, then ruled out, multiple theories as to why and how the ship was abandoned, including piracy, domestic Communist sabotage and the involvement of rum-runners. Flight 19 Flight 19 was a training flight of five TBM Avenger torpedo bombers that disappeared on December 5, 1945, while over the Atlantic. The squadron's flight plan was scheduled to take them due east from Fort Lauderdale for , north for , and then back over a final leg to complete the exercise. The flight never returned to base. The disappearance was attributed by Navy investigators to navigational error leading to the aircraft running out of fuel. One of the search and rescue aircraft deployed to look for them, a PBM Mariner with a 13-man crew, also disappeared. A tanker off the coast of Florida reported seeing an explosion and observing a widespread oil slick when fruitlessly searching for survivors. The weather was becoming stormy by the end of the incident. According to contemporaneous sources the Mariner had a history of explosions due to vapour leaks when heavily loaded with fuel, as it might have been for a potentially long search-and-rescue operation. Star Tiger and Star Ariel G-AHNP Star Tiger disappeared on January 30, 1948, on a flight from the Azores to Bermuda; G-AGRE Star Ariel disappeared on January 17, 1949, on a flight from Bermuda to Kingston, Jamaica. Both were Avro Tudor IV passenger aircraft operated by British South American Airways. Both planes were operating at the very limits of their range and the slightest error or fault in the equipment could keep them from reaching the small island. Douglas DC-3 On December 28, 1948, a Douglas DC-3 aircraft, number NC16002, disappeared while on a flight from San Juan, Puerto Rico, to Miami. No trace of the aircraft, or the 32 people on board, was ever found. A Civil Aeronautics Board investigation found there was insufficient information available on which to determine probable cause of the disappearance. Connemara IV A pleasure yacht was found adrift in the Atlantic south of Bermuda on September 26, 1955; it is usually stated in the stories (Berlitz, Winer) that the crew vanished while the yacht survived being at sea during three hurricanes. The 1955 Atlantic hurricane season shows Hurricane Ione passing nearby between 14 and 18 September, with Bermuda being affected by winds of almost gale force. In his second book on the Bermuda Triangle, Winer quoted from a letter he had received from Mr J.E. Challenor of Barbados: KC-135 Stratotankers On August 28, 1963, a pair of US Air Force KC-135 Stratotanker aircraft collided and crashed into the Atlantic west of Bermuda. Some writers say that while the two aircraft did collide there were two distinct crash sites, separated by over of water. However, Kusche's research showed that the unclassified version of the Air Force investigation report revealed that the debris field defining the second "crash site" was examined by a search and rescue ship, and found to be a mass of seaweed and driftwood tangled in an old buoy. See also List of Bermuda Triangle incidents List of topics characterized as pseudoscience Nevada Triangle Devil's Sea (or Dragon's Triangle) Sargasso Sea SS Cotopaxi Vile vortex Hurricane Alley References Citations Bibliography The incidents cited above, apart from the official documentation, come from the following works. Some incidents mentioned as having taken place within the Triangle are found only in these sources: Reprinted in paperback in 2005; . Further reading Newspaper articles ProQuest has newspaper source material for many incidents, archived in Portable Document Format (PDF). The newspapers include The New York Times, The Washington Post, and The Atlanta Constitution. To access this website, registration is required, usually through a library connected to a college or university. Flight 19 "Great Hunt On For 27 Navy Fliers Missing In Five Planes Off Florida", The New York Times, December 7, 1945. "Wide Hunt For 27 Men In Six Navy Planes", The Washington Post, December 7, 1945. "Fire Signals Seen In Area Of Lost Men", The Washington Post, December 9, 1945. SS Cotopaxi "Lloyd's posts Cotopaxi As 'Missing'", The New York Times, January 7, 1926. "Efforts To Locate Missing Ship Fail", The Washington Post, December 6, 1925. "Lighthouse Keepers Seek Missing Ship", The Washington Post, December 7, 1925. "53 On Missing Craft Are Reported Saved", The Washington Post, December 13, 1925. USS Cyclops (AC-4) "Cold High Winds Do $25,000 Damage", The Washington Post, March 11, 1918. "Collier Overdue A Month", The New York Times, April 15, 1918. "More Ships Hunt For Missing Cyclops", The New York Times, April 16, 1918. "Haven't Given Up Hope For Cyclops", The New York Times, April 17, 1918. "Collier Cyclops Is Lost; 293 Persons On Board; Enemy Blow Suspected", The Washington Post, April 15, 1918. "U.S. Consul Gottschalk Coming To Enter The War", The Washington Post, April 15, 1918. "Cyclops Skipper Teuton, 'Tis Said", The Washington Post, April 16, 1918. "Fate Of Ship Baffles", The Washington Post, April 16, 1918. "Steamer Met Gale On Cyclops' Course", The Washington Post, April 19, 1918. Carroll A. Deering "Piracy Suspected In Disappearance Of 3 American Ships", The New York Times, June 21, 1921. "Bath Owners Skeptical", The New York Times, June 22, 1921. piera antonella "Deering Skipper's Wife Caused Investigation", The New York Times, June 22, 1921. "More Ships Added To Mystery List", The New York Times, June 22, 1921. "Hunt On For Pirates", The Washington Post, June 21, 1921 "Comb Seas For Ships", The Washington Post, June 22, 1921. "Port Of Missing Ships Claims 3000 Yearly", The Washington Post, July 10, 1921. Wreckers "'Wreckreation' Was The Name Of The Game That Flourished 100 Years Ago", The New York Times, March 30, 1969. S.S. Suduffco "To Search For Missing Freighter", The New York Times, April 11, 1926. "Abandon Hope For Ship", The New York Times, April 28, 1926. Star Tiger and Star Ariel "Hope Wanes in Sea Search For 28 Aboard Lost Airliner", The New York Times, January 31, 1948. "72 Planes Search Sea For Airliner", The New York Times, January 19, 1949. DC-3 Airliner NC16002 disappearance "30-Passenger Airliner Disappears In Flight From San Juan To Miami", The New York Times, December 29, 1948. "Check Cuba Report Of Missing Airliner", The New York Times, December 30, 1948. "Airliner Hunt Extended", The New York Times, December 31, 1948. Harvey Conover and Revonoc "Search Continuing For Conover Yawl", The New York Times, January 8, 1958. "Yacht Search Goes On", The New York Times, January 9, 1958. "Yacht Search Pressed", The New York Times, January 10, 1958. "Conover Search Called Off", The New York Times, January 15, 1958. KC-135 Stratotankers "Second Area Of Debris Found In Hunt For Jets", The New York Times, August 31, 1963. "Hunt For Tanker Jets Halted", The New York Times, September 3, 1963. "Planes Debris Found In Jet Tanker Hunt", The Washington Post, August 30, 1963. B-52 Bomber (Pogo 22) "U.S.-Canada Test Of Air Defence A Success", The New York Times, October 16, 1961. "Hunt For Lost B-52 Bomber Pushed In New Area", The New York Times, October 17, 1961. "Bomber Hunt Pressed", The New York Times, October 18, 1961. "Bomber Search Continuing", The New York Times, October 19, 1961. "Hunt For Bomber Ends", The New York Times, October 20, 1961. Charter vessel Sno'Boy "Plane Hunting Boat Sights Body In Sea", The New York Times, July 7, 1963. "Search Abandoned For 40 On Vessel Lost In Caribbean", The New York Times, July 11, 1963. "Search Continues For Vessel With 55 Aboard In Caribbean", The Washington Post, July 6, 1963. "Body Found In Search For Fishing Boat", The Washington Post, July 7, 1963. SS Marine Sulphur Queen "Tanker Lost In Atlantic; 39 Aboard", The Washington Post, February 9, 1963. "Debris Sighted In Plane Search For Tanker Missing Off Florida", The New York Times, February 11, 1963. "2.5 Million Is Asked In Sea Disaster", The Washington Post, February 19, 1963. "Vanishing Of Ship Ruled A Mystery", The New York Times, April 14, 1964. "Families Of 39 Lost At Sea Begin $20-Million Suit Here", The New York Times, June 4, 1969. "10-Year Rift Over Lost Ship Near End", The New York Times, February 4, 1973. SS Sylvia L. Ossa "Ship And 37 Vanish In Bermuda Triangle On Voyage To U.S.", The New York Times, October 18, 1976. "Ship Missing In Bermuda Triangle Now Presumed To Be Lost At Sea", The New York Times, October 19, 1976. "Distress Signal Heard From American Sailor Missing For 17 Days", The New York Times, October 31, 1976. Website links The following websites have either online material that supports the popular version of the Bermuda Triangle, or documents published from official sources as part of hearings or inquiries, such as those conducted by the United States Navy or United States Coast Guard. Copies of some inquiries are not online and may have to be ordered; for example, the losses of Flight 19 or USS Cyclops can be ordered direct from the United States Naval Historical Center. Text of Feb, 1964 Argosy Magazine article by Vincent Gaddis United States Coast Guard database of selected reports and inquiries U.S. Navy Historical Center Bermuda Triangle FAQ U.S. Navy Historical C/ The Bermuda Triangle: Startling New Secrets, Sci Fi Channel documentary (November 2005) Navy Historical Center: The Loss Of Flight 19 on losses of heavy ships at sea Bermuda Shipwrecks Association of Underwater Explorers shipwreck listings page Dictionary of American Naval Fighting Ships Books Most of the works listed here are largely out of print. Copies may be obtained at your local library, or purchased used at bookstores, or through eBay or Amazon.com. These books are often the only source material for some of the incidents that have taken place within the Triangle. Into the Bermuda Triangle: Pursuing the Truth Behind the World's Greatest Mystery by Gian J. Quasar, International Marine/Ragged Mountain Press (2003) ; contains list of missing craft as researched in official records. (Reprinted in paperback (2005) ). The Bermuda Triangle, Charles Berlitz (): Out of print. The Bermuda Triangle Mystery Solved (1975). Lawrence David Kusche () Limbo Of The Lost, John Wallace Spencer () The Evidence for the Bermuda Triangle (1984), David Group () The Final Flight (2006), Tony Blackman (). This book is a work of fiction. Bermuda Shipwrecks (2000), Daniel Berg() The Devil's Triangle (1974), Richard Winer (); this book sold well over a million copies by the end of its first year; to date there have been at least 17 printings. The Devil's Triangle 2 (1975), Richard Winer () From the Devil's Triangle to the Devil's Jaw (1977), Richard Winer () Ghost Ships: True Stories of Nautical Nightmares, Hauntings, and Disasters (2000), Richard Winer () The Bermuda Triangle (1975) by Adi-Kent Thomas Jeffrey () External links Earth mysteries Geography of Miami Paranormal triangles Supernatural legends Urban legends
4988
https://en.wikipedia.org/wiki/Brezhnev%20Doctrine
Brezhnev Doctrine
The Brezhnev Doctrine was a Soviet foreign policy that proclaimed that any threat to "socialist rule" in any state of the Soviet Bloc in Central and Eastern Europe was a threat to all of them, and therefore, it justified the intervention of fellow socialist states. It was proclaimed in order to justify the Soviet-led occupation of Czechoslovakia earlier in 1968, with the overthrow of the reformist government there. The references to "socialism" meant control by the communist parties which were loyal to the Kremlin. Soviet leader Mikhail Gorbachev repudiated the doctrine in the late 1980s, as the Kremlin accepted the peaceful overthrow of Soviet rule in all its satellite countries in Eastern Europe. The policy was first and most clearly outlined by Sergei Kovalev in a September 26, 1968 Pravda article entitled "Sovereignty and the International Obligations of Socialist Countries". Leonid Brezhnev reiterated it in a speech at the Fifth Congress of the Polish United Workers' Party on November 13, 1968, which stated: "When forces that are hostile to socialism try to turn the development of some socialist country towards capitalism, it becomes not only a problem of the country concerned, but a common problem and concern of all socialist countries." This doctrine was announced to retroactively justify the invasion of Czechoslovakia in August 1968 that ended the Prague Spring, along with earlier Soviet military interventions, such as the invasion of Hungary in 1956. These interventions were meant to put an end to liberalization efforts and uprisings that had the potential to compromise Soviet hegemony inside the Soviet Bloc, which was considered by the Soviet Union to be an essential and defensive and strategic buffer in case hostilities with NATO were to break out. In practice, the policy meant that only limited independence of the satellite states' communist parties was allowed and that none would be allowed to compromise the cohesiveness of the Eastern Bloc in any way. That is, no country could leave the Warsaw Pact or disturb a ruling communist party's monopoly on power. Implicit in this doctrine was that the leadership of the Soviet Union reserved, for itself, the power to define "socialism" and "capitalism". Following the announcement of the Brezhnev Doctrine, numerous treaties were signed between the Soviet Union and its satellite states to reassert these points and to further ensure inter-state cooperation. The principles of the doctrine were so broad that the Soviets even used it to justify their military intervention in the communist (but non-Warsaw Pact) nation of Afghanistan in 1979. The Brezhnev Doctrine stayed in effect until it was ended with the Soviet reaction to the Polish crisis of 1980–1981. Mikhail Gorbachev refused to use military force when Poland held free elections in 1989 and Solidarity defeated the Polish United Workers' Party. It was superseded by the facetiously named Sinatra Doctrine in 1989, alluding to the Frank Sinatra song "My Way". The refusal to intervene in the emancipation of the Eastern European satellite states and the Pan-European Picnic then led to the fall of the Iron Curtain and the largely peaceful collapse of the Eastern Bloc. Origins 1956 Hungarian Revolution and Soviet invasion The period between 1953 and 1968 was saturated with dissidence and reformation within the Soviet satellite states. 1953 saw the death of Soviet Leader Joseph Stalin, followed closely by Nikita Khrushchev's 1956 "Secret Speech" denouncing Stalin. This denouncement of the former leader led to a period of the Soviet Era known commonly as "De-Stalinization." Under the blanket reforms of this process, Imre Nagy came to power in Hungary as the new Prime Minister, taking over for Mátyás Rákosi. Almost immediately Nagy set out on a path of reform. Police power was reduced, collectivized farms were split up and being returned to individual peasants, industry and food production shifted and religious tolerance was becoming more prominent. These reforms shocked the Hungarian Communist Party. Nagy was quickly overthrown by Rákosi in 1955, and stripped of his political livelihood. Shortly after this coup, Khrushchev signed the Belgrade Declaration which stated "separate paths to socialism were permissible within the Soviet Bloc." With hopes for serious reform just having been extinguished in Hungary, this declaration was not received well by the Hungarians. Tensions quickly mounted in Hungary with demonstrations and calls for not only the withdrawal of Soviet troops, but for a Hungarian withdrawal from the Warsaw Pact as well. By October 23 Soviet forces landed in Budapest. A chaotic and bloody suppression of revolutionary forces lasted from October 24 until November 7, ending with thousands of Hungarians murdered and many more fleeing the country. Although order was restored, tensions remained on both sides of the conflict. Hungarians resented the end of the reformation, and the Soviets wanted to avoid a similar crisis from occurring again anywhere in the Soviet Bloc. A peaceful Brezhnev Doctrine When the Hungarian Revolution of 1956 was suppressed, the Soviets adopted the mindset that governments supporting both communism and capitalism must coexist, and more importantly, build relations. This idea stressed that all people are equal, and own the right to solve the problems of their own countries themselves, and that in order for both states to peacefully coexist, neither country can exercise the right to get involved in each other's internal affairs. While this idea was brought up following the events of Hungary, they were not put into effect for a great deal of time. This is further explained in the Renunciation section. 1968 Prague Spring Notions of reform had been slowly growing in Czechoslovakia since the early-mid 1960s. However, once the Stalinist President Antonín Novotný resigned as head of the Czechoslovak Communist Party in January 1968, the Prague Spring began to take shape. Alexander Dubček replaced Novotný as head of the party, initially thought a friend to the Soviet Union. It was not long before Dubček began making serious liberal reforms. In an effort to establish what Dubček called "developed socialism", he instituted changes in Czechoslovakia to create a much more free and liberal version of the socialist state. Aspects of a market economy were implemented, travel travel restrictions were eased for citizens, state censorship loosened, the power of the secret police was limited, and steps were taken to improve relations with the west. As the reforms piled up, the Kremlin quickly grew uneasy as they hoped to not only preserve their power within Czechoslovakia, but to avoid another Hungarian-style revolution as well. Soviet panic compounded in March of 1968 when student protests erupted in Poland and Antonín Novotný resigned as the Czechoslovak President. On March 21, Yuri Andropov, the KGB Chairman, issued a grave statement concerning the reforms taking place under Dubček. "The methods and forms by which the work is progressing in Czechoslovakia remind one very much of Hungary. In this outward appearance of chaos…there is a certain order. It all began like this in Hungary also, but then came the first and second echelons, and then, finally the social democrats." Leonid Brezhnev sought clarification from Dubček on March 21, with the Politburo convened, on the situation in Czechoslovakia. Eager to avoid a similar fate as Imre Nagy, Dubček reassured Brezhnev that the reforms were totally under control and not on a similar path to those seen in 1956 in Hungary. Despite Dubček's assurances, other Soviet allies grew uneasy by the reforms taking place in an Eastern European neighbor. The First Secretary of the Ukrainian Communist Party (banned after the restoration of Ukrainian independence in 1991) called on Moscow for an immediate invasion of Czechoslovakia in order to stop Dubček's "socialism with a human face" from spreading into the Ukrainian SSR and sparking unrest. By May 6, Brezhnev condemned Dubček's system, declaring it a step toward "the complete collapse of the Warsaw Pact." After three months of negotiations, agreements, and rising tensions between Moscow and Czechoslovakia, the Soviet/Warsaw Pact invasion began on the night of August 20, 1968, which was to be met with great Czechoslovak discontent and resistance for many months into 1970. Formation of the Doctrine Brezhnev realized the need for a shift from Nikita Khrushchev's idea of "different paths to socialism" towards one that fostered a more unified vision throughout the socialist camp. "Economic integration, political consolidation, a return to ideological orthodoxy, and inter-Party cooperation became the new watchwords of Soviet bloc relations." On November 12, 1968, Brezhnev stated that "[w]hen external and internal forces hostile to socialism try to turn the development of a given socialist country in the direction of … the capitalist system ... this is no longer merely a problem for that country's people, but a common problem, the concern of all socialist countries." Brezhnev's statement at the Fifth Congress of the Polish United Workers Party effectively classified the issue of sovereignty as less important than the preservation of Soviet-style socialism. While no new doctrine had been officially announced, it was clear that Soviet intervention was imminent if Moscow perceived any satellite to be at risk of jeopardizing the Soviet hegemony. Brezhnev Doctrine in practice The vague, broad nature of the Brezhnev Doctrine allowed application to any international situation the USSR saw fit. This is clearly evident not only through the Prague Spring in 1968, and the indirect pressure on Poland from 1980 to 1981, but also in the Soviet involvement in Afghanistan starting in the 1970s. Any instance which caused the USSR to question whether or not a country was becoming a "risk to international socialism", the use of military intervention was, in Soviet eyes, not only justified, but necessary. Invasion of Afghanistan in 1979 The Soviet government's desire to link its foreign policy to the Brezhnev Doctrine was evoked again when it ordered a military intervention in Afghanistan in 1979. This was perhaps the last chapter of this doctrine's saga. In April 1978, a coup in Kabul brought the Afghan Communist Party to power with Nur Muhammad Taraki being installed as the second president of Afghanistan. The previous president, Mohammed Daoud Khan was killed during the coup. The Saur Revolution (as the coup was known) took Moscow by surprise, who preferred that the pro-Soviet Daoud Khan stay in power. The previous regime had maintained a pro-Soviet foreign policy as Daoud Khan was a Pashtun who rejected the Durand Line as the frontier with Pakistan. The Afghan Communist Party was divided into a factional struggle between factions known as the Khalq and Parcham. The Parcham was the more moderate of the two factions, arguing that Afghanistan was not ready for socialism, requiring more gradual process while the ultra-Communist Khalq favored a more radical approach. The Khalq faction was victorious and the leader of the Pacham faction Babrak Karmal fled to Moscow in fear of his life, to take up the position as Afghan ambassador in Moscow. Islamic fundamentalists took issue with the Communist party in power. As a result, a jihad was proclaimed against the Communist government. Brezhnev and other Soviet leaders falsely portrayed the United States as the one behind the jihad in Afghanistan, and the rebellion in Afghanistan was seen in Moscow not so much in the context of Afghan politics with an unpopular government pursuing policies that much of the population rejected (such as the collectivisation of agriculture), but rather in the context of the Cold War, being seen as the first stage of an alleged American plot to instigate a jihad in Soviet Central Asia where the majority of the population was Muslim. To assist the government, the Soviet Union drastically increased its military aid to Afghanistan while sending Soviet advisers to train the Afghan military. Following a split in the Communist Party, the leader of the Khalq faction, Hafizullah Amin, overthrew President Nur Muhammad Taraki and had him murdered on 8 October 1979. Soviet diplomats in Kabul had a low opinion of Taraki's ability to handle the rebellion, and an even lower one of Amin, who was regarded as a fanatic, but incompetent leader who lost control of the situation. In the fall of 1979, the leaders who pressed the most strongly for an invasion of Afghanistan to replace the incompetent Amin with Karmal, who was the man better able to preserve the communist regime's existence, were the Foreign Minister Andrei Gromyko, the Chairman of KGB, Yuri Andropov and the Defense Minister Marshal Dmitry Ustinov. The intervention was envisioned in Moscow was merely a short conflict to stabilize the situation and allow the Communist regime to rerain in power. Brezhnev was indecisive, fearing that an occupation of Afghanistan might not be the short war that Gromyko, Ustinov and Andropov kept insisting it would be, but was fearful of the possibility of an Islamic fundamentalist regime being established that would export Islam into Soviet Central Asia. As it was, the inability and unwillingness of much of the Soviet-controlled Afghan Army to fight led the Soviets to involve themselves in Afghanistan for almost 10 years. Ironically, despite what was being feared in Moscow, the United States was not supporting the Islamic fundamentalist rebellion in Afghanistan, and only started to support the mujahideen ("warriors of Allah") with weapons after the Soviet invasion, concentrating foreign policy matters in the form of linkage towards preventing Soviet expansion. During his talks with the Soviets during his time as Ambassador, Karmal coordinated with the Soviet government to replace Amin. It was this coordination that led to both Soviet soldiers and airborne units organizing a coup against the Amin-lead Afghanistan government, during which Amin was assassinated. In his place, the Soviets installed their ally, former-Ambassador Babrak Karmal, as the new lead of the government in Afghanistan. The Soviet Union, once again, fell back to the Brezhnev Doctrine for rationale, claiming that it was both morally and politically justified. It was also explained by the Soviets that they owed help to their friend and ally Babrak Karmal. Renunciation The long lasting struggle of the war in Afghanistan made the Soviets realize that their reach and influence was in fact limited. "[The war in Afghanistan] had shown that socialist internationalism and Soviet national interests were not always compatible." Tensions between the USSR and Czechoslovakia since 1968, as well as Poland in 1980, proved the inefficiencies inherent in the Brezhnev Doctrine. The Solidarity trade union protests in Poland were suppressed without outside intervention, leaving the Brezhnev doctrine effectively dead. Although the Kremlin wanted to preserve communism in its satellites, the decision was not to intervene. Gorbachev's Glasnost and Perestroika finally opened the door for Soviet Bloc countries and republics to make reforms without fear of Soviet intervention. When East Germany desperately asked for Soviet troops to put down growing unrest in 1989, Gorbachev flatly refused. Post-Brezhnev Doctrine With the agreement to terminate the Brezhnev Doctrine, later came on a new leader for the Soviets—Mikhail Gorbachev. His were much more relaxed. This is most likely due to the fact that Brezhnev Doctrine was no longer at the disposal of the Soviet Union. This had a major effect on the way that the Soviets carried out their new mentality when dealing with countries they once tried to control. This was best captured by Gorbachev's involvement with a group by the name of the Council of Mutual Economic Assistance (CMEA). This organization lessens the control that the Soviets had on all other partners of the agreement. This notion provided other countries that were once oppressed under communist intervention, to go about their own political reform. This actually carried over internally as well. In fact, the Soviet Union's biggest problem after the removal of the Brezhnev Doctrine, was the Khrushchev Dilemma. This did not address how to stop internal political reform, but how to tame the physical violence that comes along with it. It had become clear that the Soviet Union was beginning to loosen up. It is possible to pinpoint the renouncement of the Brezhnev Doctrine as to what started the end for the Soviet Union. Countries that were once micromanaged now could do what they wanted to politically, because the Soviets could no longer try to conquer where they saw fit. With that, the Soviet Union began to collapse. While the communist agenda had caused infinite problems for other countries, it was the driving force behind the Soviet Union staying together. After all, it seems that the removal of the incentive to conquer, and forcing of communism upon other nations, defeated the one thing Soviet Russia had always been about, the expansion of Communism. With the fall of the Brezhnev Doctrine, came the fall of the man, Brezhnev himself, the share of power in the Warsaw Pact, and perhaps the final moment for the Soviet Union, the fall of the Berlin Wall that had prevented the migration of East Germans to West Germany. The Brezhnev Doctrine coming to a close, was perhaps the beginning of the end for one of the strongest empires in the world's history, the Soviet Union. In other Communist countries The Soviet Union was not the only Communist country to intervene militarily in fellow countries. Vietnam deposed the Khmer Rouge in the Cambodian–Vietnamese War of 1978, which was followed by a revenge Chinese invasion of Vietnam in the Sino-Vietnamese War of 1979. Criticisms Brezhnev Doctrine as a UN violation This doctrine was even furthermore a problem in the view of the United Nations. The UN's first problem was that it permits use of force. This is a clear violation of Article 2, Chapter 4 of the United Nations Charter which states, "All members shall refrain in their international relations from the threat or use of force against the territorial integrity political independence of any state, or in any other manner inconsistent with the Purposes of the United Nations." When international law conflicts with the Charter, the Charter has precedent. It is this, that makes the Brezhnev Doctrine illegal. See also Captive Nations Imperialism Monroe Doctrine Russian-occupied territories Soviet Empire Sovietization Soviet occupations Satellite state Tankie Ulbricht Doctrine Western betrayal References Bibliography Freedman, Lawrence, and Jeffrey Michaels. "Soviet Doctrine from Brezhnev to Gorbachev." in The Evolution of Nuclear Strategy (Palgrave Macmillan, London, 2019) pp. 527–542. . Hunt, Lynn. The Making of the West: Peoples and Cultures (Bedford/St. Martin's, Boston and London. 2009). Jones, Robert A. The Soviet Concept of 'Limited Sovereignty' from Lenin to Gorbachev: The Brezhnev Doctrine (Springer, 2016). Kemp-Welch, Anthony. "Poland and the Brezhnev Doctrine (1968–1989)." Wolność i Solidarność 10 (2017): 155–223. online in English Kramer, Mark. "The Kremlin, the Prague Spring, and the Brezhnev Doctrine." in Promises of 1968 (2010) pp: 285–370. online Lesaffer, Randall. "Brezhnev Doctrine." in The Encyclopedia of Diplomacy (2018) pp: 1–5. Loth, Wilfried. "Moscow, Prague and Warsaw: Overcoming the Brezhnev Doctrine." Cold War History 1.2 (2001): 103–118. Mitchell, R. Judson. "The Brezhnev doctrine and communist ideology." The Review of Politics 34.2 (1972): 190–209. online Ouimet, Matthew: The Rise and Fall of the Brezhnev Doctrine in Soviet Foreign Policy. University of North Carolina Press, Chapel Hill and London. 2003. Pravda, September 25, 1968; translated by Novosti, Soviet press agency. Reprinted in L. S. Stavrianos, The Epic of Man (Englewood Cliffs, N.J.: Prentice Hall, 1971), pp. 465–466. Valenta, Jiri. "Soviet Decisionmaking on Afghanistan, 1979." im Soviet Decisionmaking for National Security (Routledge, 2021) pp. 218–236. External links 1968 in international relations 1968 in the Soviet Union Czechoslovakia–Soviet Union relations Eastern Bloc Foreign policy doctrines Foreign relations of the Soviet Union Leonid Brezhnev Neo-Stalinism Neo-Sovietism Prague Spring
5259
https://en.wikipedia.org/wiki/Common%20descent
Common descent
Common descent is a concept in evolutionary biology applicable when one species is the ancestor of two or more species later in time. According to modern evolutionary biology, all living beings could be descendants of a unique ancestor commonly referred to as the last universal common ancestor (LUCA) of all life on Earth. Common descent is an effect of speciation, in which multiple species derive from a single ancestral population. The more recent the ancestral population two species have in common, the more closely are they related. The most recent common ancestor of all currently living organisms is the last universal ancestor, which lived about 3.9 billion years ago. The two earliest pieces of evidence for life on Earth are graphite found to be biogenic in 3.7 billion-year-old metasedimentary rocks discovered in western Greenland and microbial mat fossils found in 3.48 billion-year-old sandstone discovered in Western Australia. All currently living organisms on Earth share a common genetic heritage, though the suggestion of substantial horizontal gene transfer during early evolution has led to questions about the monophyly (single ancestry) of life. 6,331 groups of genes common to all living animals have been identified; these may have arisen from a single common ancestor that lived 650 million years ago in the Precambrian. Universal common descent through an evolutionary process was first proposed by the British naturalist Charles Darwin in the concluding sentence of his 1859 book On the Origin of Species: History The idea that all living things (including things considered non-living by science) are related is a recurring theme in many indigenous worldviews across the world. Later on, in the 1740s, the French mathematician Pierre Louis Maupertuis arrived at the idea that all organisms had a common ancestor, and had diverged through random variation and natural selection. In Essai de cosmologie (1750), Maupertuis noted: May we not say that, in the fortuitous combination of the productions of Nature, since only those creatures could survive in whose organizations a certain degree of adaptation was present, there is nothing extraordinary in the fact that such adaptation is actually found in all these species which now exist? Chance, one might say, turned out a vast number of individuals; a small proportion of these were organized in such a manner that the animals' organs could satisfy their needs. A much greater number showed neither adaptation nor order; these last have all perished.... Thus the species which we see today are but a small part of all those that a blind destiny has produced. In 1790, the philosopher Immanuel Kant wrote in Kritik der Urteilskraft (Critique of Judgment) that the similarity of animal forms implies a common original type, and thus a common parent. In 1794, Charles Darwin's grandfather, Erasmus Darwin asked: [W]ould it be too bold to imagine, that in the great length of time, since the earth began to exist, perhaps millions of ages before the commencement of the history of mankind, would it be too bold to imagine, that all warm-blooded animals have arisen from one living filament, which endued with animality, with the power of acquiring new parts attended with new propensities, directed by irritations, sensations, volitions, and associations; and thus possessing the faculty of continuing to improve by its own inherent activity, and of delivering down those improvements by generation to its posterity, world without end? Charles Darwin's views about common descent, as expressed in On the Origin of Species, were that it was probable that there was only one progenitor for all life forms: Therefore I should infer from analogy that probably all the organic beings which have ever lived on this earth have descended from some one primordial form, into which life was first breathed. But he precedes that remark by, "Analogy would lead me one step further, namely, to the belief that all animals and plants have descended from some one prototype. But analogy may be a deceitful guide." And in the subsequent edition, he asserts rather, "We do not know all the possible transitional gradations between the simplest and the most perfect organs; it cannot be pretended that we know all the varied means of Distribution during the long lapse of years, or that we know how imperfect the Geological Record is. Grave as these several difficulties are, in my judgment they do not overthrow the theory of descent from a few created forms with subsequent modification". Common descent was widely accepted amongst the scientific community after Darwin's publication. In 1907, Vernon Kellogg commented that "practically no naturalists of position and recognized attainment doubt the theory of descent." In 2008, biologist T. Ryan Gregory noted that: No reliable observation has ever been found to contradict the general notion of common descent. It should come as no surprise, then, that the scientific community at large has accepted evolutionary descent as a historical reality since Darwin’s time and considers it among the most reliably established and fundamentally important facts in all of science. Evidence Common biochemistry All known forms of life are based on the same fundamental biochemical organization: genetic information encoded in DNA, transcribed into RNA, through the effect of protein- and RNA-enzymes, then translated into proteins by (highly similar) ribosomes, with ATP, NADPH and others as energy sources. Analysis of small sequence differences in widely shared substances such as cytochrome c further supports universal common descent. Some 23 proteins are found in all organisms, serving as enzymes carrying out core functions like DNA replication. The fact that only one such set of enzymes exists is convincing evidence of a single ancestry. 6,331 genes common to all living animals have been identified; these may have arisen from a single common ancestor that lived 650 million years ago in the Precambrian. Common genetic code The genetic code (the "translation table" according to which DNA information is translated into amino acids, and hence proteins) is nearly identical for all known lifeforms, from bacteria and archaea to animals and plants. The universality of this code is generally regarded by biologists as definitive evidence in favor of universal common descent. The way that codons (DNA triplets) are mapped to amino acids seems to be strongly optimised. Richard Egel argues that in particular the hydrophobic (non-polar) side-chains are well organised, suggesting that these enabled the earliest organisms to create peptides with water-repelling regions able to support the essential electron exchange (redox) reactions for energy transfer. Selectively neutral similarities Similarities which have no adaptive relevance cannot be explained by convergent evolution, and therefore they provide compelling support for universal common descent. Such evidence has come from two areas: amino acid sequences and DNA sequences. Proteins with the same three-dimensional structure need not have identical amino acid sequences; any irrelevant similarity between the sequences is evidence for common descent. In certain cases, there are several codons (DNA triplets) that code redundantly for the same amino acid. Since many species use the same codon at the same place to specify an amino acid that can be represented by more than one codon, that is evidence for their sharing a recent common ancestor. Had the amino acid sequences come from different ancestors, they would have been coded for by any of the redundant codons, and since the correct amino acids would already have been in place, natural selection would not have driven any change in the codons, however much time was available. Genetic drift could change the codons, but it would be extremely unlikely to make all the redundant codons in a whole sequence match exactly across multiple lineages. Similarly, shared nucleotide sequences, especially where these are apparently neutral such as the positioning of introns and pseudogenes, provide strong evidence of common ancestry. Other similarities Biologists often point to the universality of many aspects of cellular life as supportive evidence to the more compelling evidence listed above. These similarities include the energy carrier adenosine triphosphate (ATP), and the fact that all amino acids found in proteins are left-handed. It is, however, possible that these similarities resulted because of the laws of physics and chemistry - rather than through universal common descent - and therefore resulted in convergent evolution. In contrast, there is evidence for homology of the central subunits of Transmembrane ATPases throughout all living organisms, especially how the rotating elements are bound to the membrane. This supports the assumption of a LUCA as a cellular organism, although primordial membranes may have been semipermeable and evolved later to the membranes of modern bacteria, and on a second path to those of modern archaea also. Phylogenetic trees Another important piece of evidence is from detailed phylogenetic trees (i.e., "genealogic trees" of species) mapping out the proposed divisions and common ancestors of all living species. In 2010, Douglas L. Theobald published a statistical analysis of available genetic data, mapping them to phylogenetic trees, that gave "strong quantitative support, by a formal test, for the unity of life." Traditionally, these trees have been built using morphological methods, such as appearance, embryology, etc. Recently, it has been possible to construct these trees using molecular data, based on similarities and differences between genetic and protein sequences. All these methods produce essentially similar results, even though most genetic variation has no influence over external morphology. That phylogenetic trees based on different types of information agree with each other is strong evidence of a real underlying common descent. Objections Gene exchange clouds phylogenetic analysis Theobald noted that substantial horizontal gene transfer could have occurred during early evolution. Bacteria today remain capable of gene exchange between distantly-related lineages. This weakens the basic assumption of phylogenetic analysis, that similarity of genomes implies common ancestry, because sufficient gene exchange would allow lineages to share much of their genome whether or not they shared an ancestor (monophyly). This has led to questions about the single ancestry of life. However, biologists consider it very unlikely that completely unrelated proto-organisms could have exchanged genes, as their different coding mechanisms would have resulted only in garble rather than functioning systems. Later, however, many organisms all derived from a single ancestor could readily have shared genes that all worked in the same way, and it appears that they have. Convergent evolution If early organisms had been driven by the same environmental conditions to evolve similar biochemistry convergently, they might independently have acquired similar genetic sequences. Theobald's "formal test" was accordingly criticised by Takahiro Yonezawa and colleagues for not including consideration of convergence. They argued that Theobald's test was insufficient to distinguish between the competing hypotheses. Theobald has defended his method against this claim, arguing that his tests distinguish between phylogenetic structure and mere sequence similarity. Therefore, Theobald argued, his results show that "real universally conserved proteins are homologous." RNA world The possibility is mentioned, above, that all living organisms may be descended from an original single-celled organism with a DNA genome, and that this implies a single origin for life. Although such a universal common ancestor may have existed, such a complex entity is unlikely to have arisen spontaneously from non-life and thus a cell with a DNA genome cannot reasonably be regarded as the “origin” of life. To understand the “origin” of life, it has been proposed that DNA based cellular life descended from relatively simple pre-cellular self-replicating RNA molecules able to undergo natural selection (see RNA world). During the course of evolution, this RNA world was replaced by the evolutionary emergence of the DNA world. A world of independently self-replicating RNA genomes apparently no longer exists (RNA viruses are dependent on host cells with DNA genomes). Because the RNA world is apparently gone, it is not clear how scientific evidence could be brought to bear on the question of whether there was a single “origin” of life event from which all life descended. See also The Ancestor's Tale Urmetazoan Bibliography The book is available from The Complete Work of Charles Darwin Online. Retrieved 2015-11-23. Retrieved 2015-11-23. Notes References External links 29+ Evidences for Macroevolution: The Scientific Case for Common Descent from the TalkOrigins Archive. The Tree of Life Web Project Evolutionary biology Descent Last common ancestors
5429
https://en.wikipedia.org/wiki/Geography%20of%20Cambodia
Geography of Cambodia
Cambodia is a country in mainland Southeast Asia. It borders Thailand, Laos, Vietnam, the Gulf of Thailand and covers a total area of approximately . The country is situated in its entirety inside the tropical Indomalayan realm and the Indochina Time zone (ICT). Cambodia's main geographical features are the low lying Central Plain that includes the Tonlé Sap basin, the lower Mekong River flood-plains and the Bassac River plain surrounded by mountain ranges to the north, east, in the south-west and south. The central lowlands extend into Vietnam to the south-east. The south and south-west of the country constitute a long coast at the Gulf of Thailand, characterized by sizable mangrove marshes, peninsulas, sandy beaches and headlands and bays. Cambodia's territorial waters account for over 50 islands. The highest peak is Phnom Aural, sitting above sea level. The landmass is bisected by the Mekong River, which at is the longest river in Cambodia. After extensive rapids, turbulent sections and cataracts in Laos, the river enters the country at Stung Treng province, is predominantly calm and navigable during the entire year as it widens considerably in the lowlands. The Mekong's waters disperse into the surrounding wetlands of central Cambodia and strongly affect the seasonal nature of the Tonlé Sap lake. Two third of the country's population live in the lowlands, where the rich sediment deposited during the Mekong's annual flooding makes the agricultural lands highly fertile. As deforestation and over-exploitation affected Cambodia only in recent decades, forests, low mountain ranges and local eco-regions still retain much of their natural potential and although still home to the largest areas of contiguous and intact forests in mainland Southeast Asia, multiple serious environmental issues persist and accumulate, which are closely related to rapid population growth, uncontrolled globalization and inconsequential administration. The majority of the country lies within the Tropical savanna climate zone, as the coastal areas in the South and West receive noticeably more and steady rain before and during the wet season. These areas constitute the easternmost fringes of the south-west monsoon, determined to be inside the Tropical monsoon climate. Countrywide there are two seasons of relatively equal length, defined by varying precipitation as temperatures and humidity are generally high and steady throughout the entire year. Geological development Mainland Southeast Asia consists of allochthonous continental blocks from Gondwanaland. These include the South China, Indochina, Sibumasu, and West Burma blocks, which amalgamated to form the Southeast Asian continent during the Paleozoic and Mesozoic periods. The current geological structure of South China and South-East Asia is determined to be the response to the "Indo-sinian" collision in South-East Asia during the Carboniferous. The Indo-Sinian orogeny was followed by extension of the Indo-Chinese block, the formation of rift basins and thermal subsidence during the early Triassic. The Indochina continental block, which is separated from the South China Block by the Jinshajiang-Ailaoshan Suture zone, is an amalgamation of the Viet-Lao, Khorat-Kontum, Uttaradit (UTD), and Chiang Mai-West Kachin terranes, all of which are separated by suture zones or ductile shear zones. The Khorat-Kontum terrane, which includes western Laos, Cambodia and southern Vietnam, consists of the Kontum metamorphic complex, Paleozoic shallow marine deposits, upper Permian arc volcanic rocks and Mesozoic terrigenous sedimentary rocks. The central plains consist mainly of Quaternary sands, loam and clay, as most of the northern mountain regions and the coastal region are largely composed of Cretaceous granite, Triassic stones and Jurassic sandstone formations. General topography Bowl- or saucer-shaped, Cambodia covers in the south-western part of the Indochinese peninsula as its landmass and marine territory is situated entirely within the tropics. The bowl's bottom represents Cambodia's interior, about 75 percent, consisting of alluvial flood-plains of the Tonlé Sap basin, the lower Mekong River and the Bassac River plain, whose waters feed the large and almost centrally located wetlands. As humans preferably settle in these fertile and easily accessible central lowlands, major transformations and widespread cultivation through wet-rice agriculture have over the centuries shaped the landscape into distinctive regional cultivated lands. Domestic plants, such as sugar palms, Coconut trees and banana groves almost exclusively skirt extensive rice paddies, as natural vegetation is confined to elevated lands and near waterways. The Mekong traverses the north to south-east portions of the country, where the low-lying plains extend into Vietnam and reach the South China Sea at the Mekong Delta region. Cambodia's low mountain ranges - representing the walls of the bowl - remain as the result of only rather recent substantial infrastructural development and economic exploitation - in particular in remote areas - formidably forested. The country is fringed to the north by the Dangrek Mountains plateau, bordering Thailand and Laos, to the north-east by the Annamite Range, in the south-west by the Cardamom Mountains and in the South by the Elephant Mountains. Highlands to the north-east and to the east merge into the Central Highlands and the Mekong Delta lowlands of Vietnam. A heavily indented coastline at the Gulf of Thailand of length and 60 offshore islands, that dot the territorial waters and locally merge with tidal mangrove marshes - the environmental basis for a remarkable range of marine and coastal eco-regions. Soils "Sandy materials cover a large proportion of the landscape of Cambodia, on account of the siliceous sedimentary formations that underlie much of the Kingdom. Mesozoic sandstone dominates most of the basement geology in Cambodia and hence has a dominating influence on the properties of upland soils. Arenosols (sandy soils featuring very weak or no soil development) are mapped on only 1.6% of the land area." "Sandy surface textures are more prevalent than the deep sandy soils that fit the definition for Arenosols. Sandy textured profiles are common amongst the most prevalent soil groups, including Acrisols and Leptosols. The Acrisols are the most prevalent soil group occupying the lowlands - nearly half of the land area of Cambodia. Low fertility and toxic amounts of aluminium pose limitations to its agricultural use, crops that can be successfully cultivated include rubber tree, oil palm, coffee and sugar cane. The main subgroups are: Gleyic Acrisols (20.5%, Haplic Acrisols (13.3%), Plinthic Acrisol (8.7%) and Ferric Acrisol (6.3%)." Geographical extremes Northernmost point: Ta Veaeng District, Rattanakiri Province () Southernmost point: Koh Poulo Wai, Kampot Province () Easternmost point: Ou Ya Dav District, Rattanakiri Province () Westernmost point: Malai District, Banteay Meanchey Province () Regions Central plain The vast alluvial and lacustrine interconnected Cambodian flood-plain is a geologically relatively recent depression where the sediments of the Mekong and its tributaries accumulate as waters are subject to frequent course changes. The area covers . The Tonlé Sap lake and - river system occupies the lowest area. The Tonle Sap river is a waterway that branches off the Mekong near Phnom Penh in the north-westerly direction and meets the Tonle Sap lake after around . Its waters' flow reverses direction every year, caused by greatly varying amounts of water carried by the Mekong over the course of a year and the impact of monsoonal rains, that coincides with the river's maximum. The plains of the Mekong and Tonle Sap basin are confined in the North by the Dangrek and Central Annamite Mountains, and to the South by the Cardamom Mountains and Elephant Mountains. The plains completely surround the Tonle Sap Lake in the western half of the country and wind their way through the middle of the country following the course of the Mekong River. The two basins actually form a single body of water, the whole of which effects about 75% of Cambodia’s land cover. Flow reversal The Mekong river and its tributaries increase water volumes in spring (May) on the northern hemisphere, mainly caused by melting snows. As the Mekong enters Cambodia (over 95% of its waters have already joined the river) it widens and inundates large areas. The plain's deepest point - the Tonle Sap - flooded area varies from a low of around with a depth of around 1 meter at the end of the dry season (April) to and a depth of up to 9 meters in October/November. This figure rose to during 2000 when some of the worst flood conditions recorded caused over 800 deaths in Cambodia and Vietnam. Inflow starts in May/June with maximum rates of flow of around 10,000 m3/s by late August and ends in October/November, amplified by precipitation of the annual monsoon. In November the lake reaches its maximum size. The annual monsoon coincides to cease around this time of the year. As the Mekong river begins its minimum around this time of the year and its water level falls deeper than the inundated Tonle Sap lake, Tonle Sap river and surrounding wetlands, waters of the lake's basin now drains via the Tonle Sap river into the Mekong. As a result the Tonle Sap River (length around ) flows 6 months a year from South-East (Mekong) to North-West (lake) and 6 month a year in the opposite direction. The mean annual reverse flow volume in the Tonle Sap is , or about half of the maximum lake volume. A further 10% is estimated to enter the system by overland flow from the Mekong. The Mekong branches off into several arms near Phnom Penh and reaches Vietnamese territory south of Koh Thom and Loek Daek districts of Kandal Province. Southern Mountains This region represents the eastern parts of the original extent of the wet evergreen forests that cover the Cardamom - and Elephant Mountains in South-West Cambodia and along the mountains east of Bangkok in Thailand. The densely wooded hills receive rainfall of annually on their western slopes (which are subject to the South-West monsoons) but only on their eastern - rain shadow - slopes. The Cardamom/Krâvanh Mountains Occupying Koh Kong Province and Kampong Speu Province, running in a north-western to south-eastern direction and rising to more than . The highest mountain of Cambodia, Phnom Aural, at is located in Aoral District in Kampong Speu Province. The Cardamom Mountains form - including the north-western part of Chanthaburi Province, Thailand, the 'Soi Dao Mountains' - the Cardamom Mountains Moist Forests Ecoregion, that is considered to be one of the most species-rich and intact natural habitats in the region. The climate, size inaccessibility and seclusion of the mountains have allowed a rich variety of wildlife to thrive. The Cardamom and Elephant Mountains remain to be fully researched and documented. The Elephant Mountains Chuŏr Phnum Dâmrei - A north-south-trending range of high hills, an extension of the Cardamom/Krâvanh Mountains, in south-eastern Cambodia, rising to elevations of between 500 and 1,000 meters. Extending north from the Gulf of Thailand, they reach a high point in the Bok Koŭ ridge at Mount Bokor near the sea. To the south-west of the Southern mountain ranges extends a narrow coastal plain that contains the Kampong Saom Bay area and the Sihanoukville peninsula, facing the Gulf of Thailand. Northern Mountains The Dangrek Mountains A forested range of hills averaging , dividing Thailand from Cambodia, mainly formed of massive sandstone with slate and silt. A few characteristic basalt hills are located on the northern side of the mountain chain. This east–west-trending range extends from the Mekong River westward for approximately , merging with the highland area near San Kamphaeng, Thailand. Essentially the southern escarpment of the sandstone Khorat Plateau of northeastern Thailand, the Dângrêk range slopes gradually northward to the Mun River in Thailand but falls more abruptly in the south to the Cambodian plain. Its highest point is . The watershed along the escarpment in general terms marks the boundary between Thailand and Cambodia, however there are exceptions. The region is covered in dry evergreen forest, mixed dipterocarp forest, and deciduous dipterocarp forests. Tree species like Pterocarpus macrocarpus, Shorea siamensis and Xylia xylocarpa var. kerrii dominate. Illegal logging are issues on both, the Thai as well as on the Cambodian side, leaving large hill stretches denuded, vulnerable tree species such as Dalbergia cochinchinensis have been affected. Forest fires are common during the dry season. Annamite Range Lying to the east of the Mekong River, the long chain of mountains called the Annamite Mountains of Indochina and the lowlands that surround them make up the Greater Annamites ecoregion. Levels of rainfall vary from annually. Mean annual temperatures are about . This eco-region contains some of the last relatively intact moist forests in Indochina. Moisture-laden monsoon winds, that blow in from the Gulf of Tonkin ensure permanent high air humidity. Plants and animals adapted to moist conditions, to seek refuge here and evolve into highly specialized types that are found nowhere else on Earth. Ethnically diverse More than 30 ethnic groups of indigenous people live in the Annamites, each with their distinctive and traditional music, language, dress and customs. The natural resources of the Greater Annamites are vital to all of these people. Eastern Highlands Tall grasses and deciduous forests cover the ground east of the Mekong River in Mondulkiri, where the transitional plains merge with the eastern highlands at altitudes from . The landscape has suffered from rubber farming, logging and particularly mining, although sizable areas of pristine jungle survive, which are home to rare and endemic wildlife. Coast Cambodia's coastal area covers , distributed among four provinces: Sihanoukville province, Kampot province, Koh Kong province, and Kep province. The total length of the Cambodian coastal area has been disputed. The most widely accepted length is , a 1997 survey by the DANIDA organization announced a length at , and in 1973 the Oil Authority found the coast to be long. The Food and Agriculture Organization claims a length of in one of its studies. The southern mountain ranges drain to the south and west towards the shallow sea. Sediments on the continental shelf are the basis for extensive mangroves marshes, in particular in the Koh Kong province and the Ream National Park. Islands Cambodia’s islands fall under administration of the 4 coastal provinces. "There are 60 islands in Cambodia's coastal waters. They include 23 in Koh Kong province, 2 in Kampot province, 22 in Sihanoukville and 13 in Kep city.[sic]" Most islands are, apart from the two small groups of the outer islands, in relative proximity to the coast. The islands and the coastal region of Koh Kong Province are mainly composed of upper Jurassic and lower Cretaceous sandstone massives. The north-westernmost islands near and around the Kaoh Pao river delta (Prek Kaoh Pao) area are to a great extent sediments of estuaries and rivers, very flat and engulfed in contiguous mangrove marshes. Climate Cambodia's climate, like that of much of the rest of mainland Southeast Asia is dominated by monsoons, which are known as tropical wet and dry because of the distinctly marked seasonal differences. The monsoonal air-flows are caused by annual alternating high pressure and low pressure over the Central Asian landmass. In summer, moisture-laden air—the southwest monsoon—is drawn landward from the Indian Ocean. The flow is reversed during the winter, and the northeast monsoon sends back dry air. The southwest monsoon brings the rainy season from mid-May to mid-September or to early October, and the northeast monsoon flow of drier and cooler air lasts from early November to March. Temperatures are fairly uniform throughout the Tonlé Sap Basin area, with only small variations from the average annual mean of around . The maximum mean is about ; the minimum mean, about . Maximum temperatures of higher than , however, are common and, just before the start of the rainy season, they may rise to more than . Minimum night temperatures sporadically fall below . in January, the coldest month. May is the warmest month - although strongly influenced by the beginning of the wet season, as the area constitutes the easternmost fringe of the south-west monsoon. Tropical cyclones only rarely cause damage in Cambodia. The total annual rainfall average is between , and the heaviest amounts fall in the southeast. Rainfall from April to September in the Tonlé Sap Basin-Mekong Lowlands area averages annually, but the amount varies considerably from year to year. Rainfall around the basin increases with elevation. It is heaviest in the mountains along the coast in the southwest, which receive from to more than of precipitation annually as the southwest monsoon reaches the coast. This area of greatest rainfall drains mostly to the sea; only a small quantity goes into the rivers flowing into the basin. Relative humidity is high throughout the entire year; usually exceeding 90%. During the dry season daytime humidity rates average around 50 percent or slightly lower, climbing to about 90% during the rainy season. Hydrology The Mekong River and its tributaries comprise one of the largest river systems in the world. The central Tonle Sap, the Great Lake has several input rivers, the most important being the Tonle Sap River, which contributes 62% of the total water supply during the rainy season. Direct rainfall on the lake and the other rivers in the sub-basin contribute the remaining 38%. Major rivers are the Sen river, Sreng River, Stung Pouthisat River, Sisophon River, Mongkol Borei River, and Sangkae River. Smaller rivers in the southeast, the Cardamom Mountains and Elephant Range form separate drainage divides. To the east the rivers flow into the Tonle Sap, as in the south-west rivers flow into the Gulf of Thailand. Toward the southern slopes of the Elephant Mountains, small rivers flow south-eastward on the eastern side of the divide. The Mekong River flows southward from the Cambodia-Laos border to a point south of Kratié (town), where it turns west for about and then turns southwest towards Phnom Penh. Extensive rapids run north of Kratie city. From Kampong Cham Province the gradient slopes very gently, and inundation of areas along the river occurs at flood stage. From June through November—through breaks in the natural levees that have built up along its course. At Phnom Penh four major water courses meet at a point called the Chattomukh (Four Faces). The Mekong River flows in from the northeast and the Tonle Sap river emanates from the Tonle Sap—flows in from the northwest. They divide into two parallel channels, the Mekong River proper and the Bassac River, and flow independently through the delta areas of Cambodia and Vietnam to the South China Sea. The flow of water into the Tonle Sap is seasonal. In spring, the flow of the Mekong River, fed by monsoon rains, increases to a point where its outlets through the delta can't handle the enormous volume of water. At this point, the water pushes northward up the Tonle Sap river and empties into the Tonle Sap lake, thereby increasing the size of the lake from about to about at the height of the flooding. After the Mekong's waters crest — when its downstream channels can handle the volume of water — the flow reverses, and water flows out of the engorged lake. As the level of the Tonle Sap retreats, it deposits a new layer of sediment. The annual flooding, combined with poor drainage immediately around the lake, transforms the surrounding area into marshlands, unusable for agricultural purposes during the dry season. The sediment deposited into the lake during the Mekong's flood stage appears to be greater than the quantity carried away later by the Tonle Sap River. Gradual silting of the lake would seem to be occurring; during low-water level, it is only about deep, while at flood stage it is between deep. Vegetation & ecoregions Cambodia has one of the highest levels of forest cover in the region as the interdependence of Cambodia’s geography and hydrology makes it rich in natural resources and biological diversity - among the bio-richest countries in Southeast Asia. The Royal Government of Cambodia estimates Cambodia contains approximately 10.36 million hectares of forest cover, representing approximately 57.07% of Cambodia’s land area (2011). On the contrary, international observers and independent sources provide rather different numbers. Consensus permeates, as most sources agree, that deforestation in Cambodia, loss of seasonal wetlands and habitat destruction - among countless minor factors - correlates with the absence of strict administrative control and indifference in law enforcement - not only in Cambodia but the entire region. Figures and assessments are numerous as are available sources. as seen in numbers below, which provide a wide range for interpretation. About (1%) of forest cover is planted forest. Overall Cambodia’s forests contain an estimated 464 million metric tonnes of carbon stock in living forest biomass. Approximately 40% of Cambodia’s Forests have some level of protection, while one of the Cambodia Millennium Development Goals targets is to achieve a 60% forest cover by 2015. According to the Forestry Administration statistics, a total of 380,000 hectares of forest were cleared between 2002 and 2005/2006 - a deforestation rate of 0.5% per year. The main cause of deforestation has been determined to be large-scale agricultural expansions. Southern Annamites Montane Rain Forests ecoregion The Southern Annamites Montane Rain Forests ecoregion of the montane forests of Kontuey Nea, "the dragon's tail" in the remote north-west of Cambodia, where the boundaries of Cambodia, Laos, and Vietnam meet [this is in the northeast, not the northwest?], is remarkably rich in biodiversity. The relatively intact forests occupy a broad topographic range - from lowlands with wet evergreen forests to montane habitats with evergreen hardwood and conifer forests. The complex geological, topographic and climatic ( rainfall and temperature ) facets that characterize the region make forest structure and composition unique and very variable. There is an unusually high number of near-endemic and endemic species among the many species to be found in the area. The entire eco-region has a size of . Teri maa ka bhosda. The Great Lake ecosystem The Tonle Sap, also known as the Great Lake in central Cambodia is the largest freshwater lake in Southeast Asia and one of the richest inland fishing grounds in the world. The Lake functions as a natural flood water reservoir for the Mekong system as a whole and therefore is an important source of water for the Mekong Delta during the dry season. The ecosystem has developed as a result of the Mekong’s seasonal flow fluctuations. A belt of freshwater mangroves known as the "flooded forest" surrounds the lake. The floodplains in turn are surrounded by low hills, covered with evergreen seasonal tropical forest with substantial dipterocarp vegetation or deciduous dry forest. The eco-region consists of a mosaic of habitats for a great number of species. The forest gradually yields to bushes and finally grassland with increasing distance from the lake. Henri Mouhot: "Travels in the Central Parts of Indo-China" 1864 On higher quality soils or at higher elevation, areas of mixed deciduous forest and semi-evergreen forests occur. This variety of vegetation types accounts for the quantity and diversity of species of the Great Lake ecosystem. Interlocking forest, - grassland and marshland patches provide the many facets and refugia for the abundant local wildlife. The lake’s flooded forest and the surrounding floodplains are of utmost importance for Cambodia's agriculture as the region represents the cultural heart of Cambodia, the center of the national freshwater fishery industry - the nation's primary protein source. Threats to the lake include widespread pollution, stress through growth of the local population which is dependent on the lake for subsistence and livelihood, over-harvesting of fish and other aquatic - often endangered - species, habitat destruction and potential changes in the hydrology, such as the construction and operation of dams, that disrupt the lake's natural flood cycle. However, concerns that the lake is rapidly filling with sediment seem - according to studies - to be unfounded at the present time. Wetlands Wetlands cover more than 30% of Cambodia. In addition to the Mekong River and the Tonle Sap floodplain there are the Stung Sen River and the coastal Stung Koh Pao - and Stung Kep estuaries of Koh Kong Province and Kep Province. The freshwater wetlands of Cambodia represent one of the most diverse ecosystems worldwide. The area’s extensive wetland habitats are the product of the annual Mekong maximum, the simultaneous wet season and the drainage paths of a number of minor rivers. See also:Geography of Cambodia#Hydrology The numerous and varied wetlands are Cambodia's central and traditional settlement area, the productive environments for rice cultivation, freshwater fisheries, other forms of agriculture and aquaculture and the constantly growing tourism sector. Considering the eco-region's importance, a variety of plans for local wetland management consolidation exist with varying degrees of completion. Coastal habitats The Cambodian coastline consists of of over 30 species of mangroves - among the most biologically diverse wetlands on earth. The most pristine mangrove forests are found in Koh Kong Province. In addition to mangroves, sea-grass beds extend throughout the coastal areas, especially in Kampot Province, the Sihanoukville Bay Delta and the Kep municipal waters. The meadows are highly productive, but few animals feed directly on the grasses. Those that do tend to be vertebrates such as sea turtles, dabbling ducks and geese. "With their roots deep in mud, jagged and gnarled mangrove trees are able to grow in the brackish wetlands between land and sea where other plant life cannot survive. The trees offer refuge and nursery grounds for fish, crabs, shrimp, and mollusks. They are nesting - and migratory sites for hundreds of bird species. They also provide homes for monkeys, lizards, sea turtles, and many other animals as well as countless insects." "Until relatively recently, the mangroves of Koh Kong, Cambodia have remained relatively intact. This is partly because of the region’s location — it is an isolated, inaccessible place — and because decades of war and conflict perversely protected the forests from over-exploitation. Local people, however, tended to use the forest's sustainability, for food, fuel, medicine, building materials, and other basic needs." Fauna Cambodia is home to a wide array of wildlife. There are 212 mammal species, 536 bird species, 176 reptile species (including 89 subspecies), 850 freshwater fish species (Tonlé Sap Lake area), and 435 marine fish species. Many of the country's species are recognized by the IUCN or World Conservation Union as threatened, endangered, or critically endangered due to deforestation and habitat destruction, poaching, illegal wildlife trade, farming, fishing, and unauthorized forestry concessions. Intensive poaching may have already driven Cambodia's national animal, the Kouprey, to extinction. Wild tigers, Eld's deer, wild water buffaloes and hog deer are at critically low numbers. Protected areas "The 1993 Royal Decree on the Protection of Natural Areas recognized 23 protected areas, which at the time covered more than 18% of the country’s total land area." Natural parks (sometimes described as ‘national parks’) Wildlife reserves Protected scenic view areas (sometimes described as ‘protected landscapes’) Multi-purpose areas Political and human geography Cambodia borders Vietnam over a length of , Thailand over a length of and Laos over a length of , with in total and an additional of coastline. The capital (reach thani) and provinces (khaet) of Cambodia are first-level administrative divisions. Cambodia is divided into 25 provinces including the capital. Municipalities and districts are the second-level administrative divisions of Cambodia. The provinces are subdivided into 159 districts and 26 municipalities. The districts and municipalities in turn are further divided into communes (khum) and quarters (sangkat). Land use Cambodia, Laos and Vietnam have experienced major changes in land use and land cover over the last two decades. The emergence from cold war rivalries and recent major economic reforms result in a shift from subsistence agrarian modes of production to market-based agricultural production and industrialized economies, which are heavily integrated into regional and global trade systems. Regional divisions Cambodia's boundaries were for the most part based upon those recognized by France and by neighboring countries during the colonial period. The boundary with Thailand runs along the watershed of the Dangrek Mountains, although only in its northern sector. The border with Laos and the border with Vietnam result from French administrative decisions and do not follow major natural features. Border disputes have broken out in the past and do persist between Cambodia and Thailand as well as between Cambodia and Vietnam. Area and boundaries Area: total: land: water: Maritime claims: territorial sea: contiguous zone: exclusive economic zone: continental shelf: Elevation extremes: lowest point: Gulf of Thailand highest point: Phnum Aoral Border disputes Cambodian–Thai border dispute Cambodian–Vietnamese land dispute Lakes Tonlé Sap Lake Yak Loum Crater Lake – Ratanakiri Natural resources Oil and natural gas - In addition to the four parts of mining project, the oilfield, Block A was discovered in 2005 and located offshore in the gulf of Thailand Chevron would operate and hold a 30% interest Block A which cover . It is expected to get 30-year-production permit in the second quarter of 2011. In late 1969, the Cambodian government granted a permit to a French company to explore for petroleum in the Gulf of Thailand. By 1972 none had been located, and exploration ceased when the Khmer Republic (see Appendix B) fell in 1975. Subsequent oil and gas discoveries in the Gulf of Thailand and in the South China Sea, however, could spark renewed interest in Cambodia's offshore area, especially because the country is on the same continental shelf as its Southeast Asian oil-producing neighbors. Timber Dipterocarpus alatus (chheuteal tan) Sawnwood, veneer, plywood Anisoptera glabra (mersawa, phdiek) Sawnwood, veneer, plywood Hopea odorata (koki) Sawmilling, construction (bridges, boats) Shorea vulgaris (choë(r) chông) Sawmilling, construction (housing) Tarrietia javanica Sawnwood (decorative, furniture) Gemstones - Gemstone areas are located in Samlot district of Battambang, Paillin, Ratanakkiri, and Takéo Province Iron ore - Hermatite (Fe2O3); Magnetite (Fe3O4); Limonite (2Fe2O3, 3H2O) - was found in two areas, one located in Phnom Deck and the others located in Koh Keo of Preah Vihear Province, and Thalaborivath of Stung Treng Province. According to General Department of Mineral, the total iron reserves in Phnom Deck area are estimated at 5 to 6 Million tons and other deposits may add 2 to 3 Million tons. Gold - Gold deposit was found in four provinces: Kampong Cham (The Rumchek in Memot area), Kampong Thom (Phnom Chi area), Preah Vihear (Phnom Deck in Roveing district), Ratanakiri (Oyadav district) and Mondulkiri Bauxite – was found in Battambang Province and Chhlong district in Mondulkiri Province. Antimony (Sb) – found in Sre Peang area, Pursat Province Chromium (Cr) – found in Sre Peang area, Pursat Province manganese phosphates Hydro-power - Hydroelectric dams: Lower Se San 2 Dam, Stung Treng Dam Arable land Marine resources Total renewable water resources: (2011) Freshwater withdrawal (domestic/industrial/agricultural): Total: /yr (4%/2%/94%) Per capita: /yr (2006) Environmental issues Natural hazards Monsoonal rains (June to November) Mekong flooding Occasional droughts Human impact Issues Illegal logging activities throughout the country rubber tree mono-cultures and strip mining for gold in the eastern highlands gem mining in the western region along the border with Thailand destruction of mangrove swamps threatens natural fisheries, illegal fishing and over-fishing large scale sand mining in river beds and estuaries of Koh Kong's mangrove marshes affects tidal balance A nascent environmental movement has been noticed by NGO's - and it is gaining strength, as the example of local resistance against the building of a Chinese hydro-electric dam in the Areng Valley shows. Cambodia has a bad but improving performance in the global Environmental Performance Index (EPI) with an overall ranking of 146 out of 180 countries in 2016. This is among the worst in the Southeast Asian region, only ahead of Laos and Myanmar. The EPI was established in 2001 by the World Economic Forum as a global gauge to measure how well individual countries perform in implementing the United Nations' Sustainable Development Goals. The environmental areas where Cambodia performs worst on the EPI (i.e. highest ranking) are air quality (148), water resource management (140) and health impacts of environmental issues (137), with the areas of sanitation, environmental impacts of fisheries and forest management following closely. Cambodia has an unusually large expanse of protected areas, both on land and at sea, with the land-based protections covering about 20% of the country. This secures Cambodia a better than average ranking of 61 in relation to biodiversity and habitat, despite the fact deforestation, illegal logging, construction and poaching are heavily deteriorating these protections and habitats in reality, partly fueled by the government's placement of economic land concessions and plantations within protected areas. In November 2017, the U.S. cut funds to help clear unexploded ordnance including land mines and chemical weapons in Cambodia which it had dropped during the Vietnam War. Consequences Flooding Deforestation Soil erosion in rural areas Declining fish stocks Decreasing access to clean water Habitat loss and declining biodiversity International agreements and conventions Cambodia is party to the following treaties: Convention on Biological Diversity Convention on Climate Change MARPOL 73/78 Tropical Timber 94 Ramsar Convention on Wetlands Signed, but not ratified: Law of the Sea See also References External links National Ministry of Land Management Urban Planning and Construction Forestry Administration Law on Forestry law on land use Ministry of Water Resources and Meteorology Tonle Sap Authority Economic Land Concession Environmental Law An Assessment of Cambodia’s Draft Environmental Impact Assessment Law Climate Change Department International National Library of France National Aquaculture Legislation Cambodia Forestry Outlook Study FAO UN Mekong River Commission National Adaptation Programme of Action to Climate Change (NAPA) World reference base for soil resources Further reading Stories from the Mekong Cardamoms 'one of the crown jewels... Kampot's forgotten Karst formations
5581
https://en.wikipedia.org/wiki/Armed%20Forces%20of%20Croatia
Armed Forces of Croatia
The Armed Forces of the Republic of Croatia () is the military service of Croatia. The President is the Armed Forces Commander-in-Chief, and exercises administrative powers in times of war by giving orders to the chief of staff, while administration and defence policy execution in peacetime is carried out by the Government through the Ministry of Defence. This unified institution consists of land, sea, and air branches referred to as: Croatian Army (Hrvatska kopnena vojska - HKoV) Croatian Navy (Hrvatska ratna mornarica - HRM) Croatian Air Force (Hrvatsko ratno zrakoplovstvo - HRZ) The Croatian Armed Forces are charged with protecting the Republic as well as supporting international peacekeeping efforts, when mandated by the NATO, United Nations and/or European Union. The Army has 650 AFVs, around 150 pieces of artillery, 100 MLRSs, around 75 tanks, and 25 SPGs. The Air Force has 12 MiG-21 jet fighters, 10 combat-transport Mi-171 and 16 OH-58 attack helicopters. The Navy has 30 ships, out of which five 60-80 metre fast attack craft are used in offensive capabilities. Strength In 2020, Armed Forces had 15,605 members, of which 14,325 were active military personnel and 1,280 civil servants. Of the 14,506 active military personnel, 3,033 were officers, 5,549 non-commissioned officers, 5,214 soldiers, 529 military specialists, 307 civil servants and 973 other employees. Total available male manpower aged 16–49 numbers 1,035,712, of which 771,323 are technically fit for military service. Male citizens are now no longer subject to compulsory military service since January 1, 2008. However, the last generation of 2007 servicemen was also absolved of compulsory service by an act from then Minister of Defence Berislav Rončević. Budget The trajectory of Croatian military budget and spending was constantly below 2% of GDP, a major difference from the 1990s when defence expenditure represented a major stake in Croatian budgetary expenditure due to then ongoing Croatian War of Independence. For example, 1995 Croatian defence budget stood at 12.4 billion Croatian Kuna or around 10% of GDP, which was also represented at the time highest defence expenditure rate. In late 2019, Croatian Government issued revised defence expenditure strategy which will see country increase its defence expenditure to gradually meet 2% NATO target, with 2019 and 2020 defence budgets seeing immediate revisions and increases to meet new spending plan. Defence expenditure in 2025 therefore based on current projections could reach €1.75 billion or around 2% of GDP which would meat NATO recommendation. However, if defence pensions are included in Croatia's defence expenditure, then Croatia already meets 2% target recommended by the NATO. Some €1140 million was paid in defence pensions to some 97000 individuals in Croatia. Defence expenditures in recent years (source Croatian MOD); Croatia adopted € and therefore, Kuna is no longer national currency in circulation or used in statistics. As of January 1, 2023 exchange rate for €1 = 7.53 Kuna. (*) - projected expenditure (**) - proposed expenditure Although the budget has been decreased in percentage of BDP from year to year, overall budget increased and the Croatian Armed Forces were able to maintain military readiness and to participate in major NATO exercises in Croatia and overseas. Downsizing of the armed forces has allowed for more funds to be allocated to modernization over the past few years with an average of 1.6 billion HRK spent on modernization, infrastructure and construction of new facilities. A $3 billion modernization plan was proposed by the then Prime Minister Ivica Račan of the SDP led government in 2003., with planned modernization starting in 2006. and ending in 2015.. However it has been delayed due to the subsequent economic recession that has cost the Croatian MOD several billion HRK since 2006.. A new plan under former Prime Minister Zoran Milanović should define exactly how and what the Croatian armed forces should look like by 2023.. A defense white paper was published in 2015 with emphasis placed on modernization of the Army. Dr. Franjo Tuđman Military Academy The Dr. Franjo Tuđman Military Academy, named after Franjo Tuđman, acts as a school of higher learning responsible for training and educating future generations of military personnel. The academy consists of several schools including "Ban Josip Jelačić", "Blago Zadro", "Katarina Zrinska", the Officers Academy, and a school for non commissioned officers. The academy has 300 full-time staff and is the only military academy in Croatia. Each year also 100–120 foreign nationals attend the academy. Commander The Commander-in-Chief of all Croatian armed forces in peace and war is the President of the Republic. The Commander-in-Chief prescribes the organisation of the Croatian Armed Forces at the proposal of the Chief of General Staff, with consent of the Minister of Defence. The Armed Forces consist of peacetime and wartime components. The peacetime component is composed of the active military officers, civil servants and employees in the Croatian Armed Forces, cadets, and conscripts serving a 6-month national service and reservists when on military exercise. The wartime component of the Armed Forces includes all other reservists. The General Staff is part of the Ministry of Defence in charge of commanding, training and use of the Armed Forces. It also has a number of units under its direct command, including the Special Operations Battalion, Honour Guard Battalion and several others. In peace, the Commander-in-Chief exercises his command through the Minister of Defence. In war and in cases where the Minister of Defence is not fulfilling his orders, the Commander-in-Chief exercises his command directly through the General Staff Commander. The Croatian Parliament exercises democratic control over the Armed Forces by adopting defence strategy, defence budget, and defence laws. Special Forces Command Special Forces Command (Zapovjedništvo specijalnih snaga, ZSS) was established in February 2015, succeeding the Special Operations Battalion (Croatia), in accordance with the Long-term Development Plan of the Croatian Armed Forces in the period 2015–2024. The command staff is composed of the members who served in the special units, guards brigades and reconnaissance units of the Croatian Armed Forces. The main mission of the Special Forces Command is to ensure combat readiness of the special forces in the protection of the territorial integrity, sovereignty and independence of the Republic of Croatia, as well as for the participation in NATO-led operations. Colonel Perica Turalija is the current commanding officer of the command. The Croatian General Staff exercises direct command over the battalion which thus elevates the unit to strategic level deployment for quicker reaction and overall better and faster response to tactical and strategic situations. Also, this means that members of all three branches of the Croatian armed forces can apply for selection. Other special operations units are the Military Intelligence Battalion (Vojno-obavještajna bojna, VOB) and Special Military Police Company (Satnija specijalne vojne policije, SS VP). The duties of an Honor Guard are performed by the Honor Guard Battalion (Počasno zaštitna bojna), located at Zagreb in the Tuškanac military base. Projects A long term modernization plan for 2015–2024 has been published outlining overall goals and is available for download (102 pages) at the Ministry of Defense of Croatia website. According to earlier reports from the government, the Croatian Armed Forces are set to receive vitally needed new equipment. Army Modernisation of M-84A and M-84A4 Snajper MBTs and upgrade to M-84D standard. The programme is at a standstill and is unlikely to receive further funding due to high costs involved of nearly 20 million kuna per tank, but overhaul of existing fleet might be completed by the end of 2017 at cost of 120 million kuna (24 vehicles already overhauled at cost of 60 million kuna). The Croatian Army is looking at replacing 128 M80A IFV in its inventory, Croatian Army requires 108 Vehicles, 88 Infantry combat vehicles, 4 Driver training vehicles, 8 Armoured Ambulance vehicles and 8 command vehicles. It's announced that M2 Bradley in M2A2 ODS variant was chosen, 60 + 24 vehicles for spare parts will be delivered in 2020 as a donation from United States and modernised in Croatia by BAE Systems.In 2022 Croatia bought 89 M2 Bradley Vehicles. Talks are being held with US and German governments on possible purchase of M270 Multiple Launch Rocket System, Croatian requirements call for initially 8 systems to form a two battery teams at independent Artillery regiment, with potential for further 16 systems to equip 2 active brigades for total of 24 systems + simulators and training aids. Number of systems Croatian Army might purchase will solely depend on price of the entire purchase and delivery dates, Croatian Defence budget has set aside some 200 million kuna or $32 million for this programme, although it is likely said systems might be donated by the US for symbolic price in turn Croatian MOD paying only VAT. The initial order of 8 Elbit UT30Mk2 CRO unmanned turrets for Patria AMV, armed with 30mm gun and Spike launchers in 2019, value of the contract: $14.9 million. Scores of smaller programmes, communication equipment, night vision capability, electronic sensors, NBC equipment, battlefield management systems and modernisation of artillery systems with new sights and electronic fire control systems are planned. Procurement of advanced short to medium range NATO SAM system. As of now no real funding has been mentioned other than the statement that this project is a priority and current requirements call for one battery. Estimated value – $50–70 million for new system. Alternately a second-hand system might be obtained as a donation by the United States, in which case only the VAT cost will be incurred. Procurement of short range SAM systems – no indication of what system or specifications other than a requirement for a range of up to 10 km. Intent is for up to 3 batteries with one battery being ready by 2015/6. Total funding for this programme has not been made public yet, but similar western systems tend to be in range of $17–20 million per battery. Air Force Croatia has a requirement for a small number of fighter aircraft (6-12) to replace the MiG-21. A plan to acquire ex-Israeli F-16C/Ds was cancelled in January 2019 after the United States refused to allow Israel to sell the aircraft. The Croatian interdepartmental commission for the procurement of multi-purpose combat aircraft has sent requests for proposals (RFP) for two new fighter types (F-16 Block 70/72 and JAS-39C/D) and five second hand fighters types (F-16AM/BM, F-16C/D, F-16 Barak, Eurofighter Typhoon and Dassault Rafale). Croatia's government said it expects to receive the new offers by May and make the decision by August 2020. In 2021 Croatia bought 12 used Dassault Rafale from France. Purchase of up to 10–12 transport helicopters (after 2020) – replacing older Mi-8. With the Government indicating that purchase of 10 new helicopters might be the only option after 2020. Most likely candidate is the UH-60M since the United States donated two new helicopters and Croatia decided to purchase additional two. Programme cost: 2.5 billion kuna. Navy Navy plans are still being worked on but present plans call for a moderate expansion of the naval force. 5 new patrol boats, locally built, 42 metres in length. Cost of programme 750 million kuna, or 375 million for first 5 ships, first to enter service in 2018. 2 new corvettes – 80–125 metres in length. Cost of programme 3.0 billion kuna. Programme is at standstill due to lack of funds, feasible only after 2020. Overhaul of existing 2 Kralj class fast attack crafts, including new engines. Cost of programme – 40 million kuna. Overhaul of sea radar Falcon 2 Enhanced Peregrine – programme is being financed by US government at estimated cost of $8 million. Possible purchase of 2nd minesweeper before 2020; although there are only indications that this might happen if funds can be allocated. Programme cost – 80 million kuna. Arms exports As a small country, Croatia has a relatively well developed arms industry that is highly competitive internationally with significant annual arms exports. In 2012, Croatia managed to export nearly €120 million. However it has been reported in The New York Times that Croatia has been arming Syrian rebels with Croatian manufactured arms used during the Homeland War, arms Croatia no longer uses due to their obsolescence. Nevertheless, these arms played a crucial role in some significant rebel gains during 2012. As a result of these arms sales into this volatile region the Croatian government ordered the immediate withdrawal of the Croatian UN Golan Heights contingent to avoid their being targeted in retaliation. In 2013 Croatia exported €143 million worth of arms, however it is not clear if this also includes $36.5 million worth of arms Croatia exported to Jordan for Syrian rebels. Croatia was the top supplier of arms to Syrian rebels in 2013, but much of it through illicit channels without Croatian government approval or knowledge. Most of these arms were exported via Jordan. In 2014 Croatian arms exports reached 1.5 bn HRK (Croatian kuna) or €200 million or $257 million, the majority of exports being to NATO allies and Australia. In late 2014 the Croatian Defence Minister announced a major export deal to Iraq including the State of Kurdistan. This agreement includes the sale of 20,000 VHS Rifles, 150,000 complete sets of uniforms, helmets and associated equipment valued at €100 million. Croatian arms exports are growing steadily at 10–15% year-on-year and were expected to reach 1.75 billion HRK in 2015 or around €230 million, although much of the equipment exported is non-lethal. Croatian firms are well positioned on some major arms tenders in the Middle East, supplying complex military hardware such as the Patria AMV incorporating a newly developed 30mm overhead weapon station (valued at €1.25 million each) and said vehicles valued at €1.75 million. Kuwait, the UAE, and Saudi Arabia have been mentioned as potential customers, although no concrete contracts have been signed so far. Croatian firms are participating in Kuwaiti and UAE tenders for next-generation APC programmes, each valued at billions of euros. Croatian arms exports have grown steadily for the better part of this decade and have reached €325 million per year, placing Croatia in the top 10 arms exporters within NATO, behind the US, Germany, UK, France, Netherlands, Spain, Italy, Poland and Norway. The vast majority of these exports are to NATO partners such as the US, Norway, Australia, Canada, France. Croatia granted €5.75 billion in export licenses in 2016 and 2017; however, only a fraction of this sum has materialised in actual arms exports. International cooperation On April 1, 2009 Croatia joined NATO and on July 1, 2013, it became the 28th member of the European Union. The Croatian Armed Forces participate in many of the (military) aspects of both organisations as well as actively participating in many United Nations peacekeeping operations worldwide. Gallery See also Croatian military ranks Croatian War of Independence List of Croatian soldiers Military history of Croatia Military Security and Intelligence Agency Croatian Ministry of Defence Orders, decorations, and medals of Croatia Citations References External links Croatian Armed Forces Official website Croatian Forces International Volunteers Association official website Croatian Military Academy official website - Defense planning and procurement. - long term planning and long term defense strategy Photos BSD CROMIL - Croatian military's official English-language magazine Permanent Structured Cooperation
5723
https://en.wikipedia.org/wiki/Constellations%20%28journal%29
Constellations (journal)
Constellations: An International Journal of Critical and Democratic Theory is a quarterly peer-reviewed academic journal of critical and democratic theory and successor of Praxis International. It is currently edited by Jean L. Cohen, Cory Pope, and Hubertus Buchstein. Ertug Tombus is the managing editor of the journal since 2010. Seyla Benhabib, Nancy Fraser and Andrew Arato are the co-founding former editors. With an international editorial contribution, it is based at the New School in New York. Nadia Urbinati, Amy Allen, and Andreas Kalyvas are former co-editors. References External links Sociology journals Academic journals established in 1994 Quarterly journals Wiley-Blackwell academic journals English-language journals
5884
https://en.wikipedia.org/wiki/Charles%20Dickens
Charles Dickens
Charles John Huffam Dickens (; 7 February 1812 – 9 June 1870) was an English writer and social critic who created some of the world's best-known fictional characters and is regarded by many as the greatest novelist of the Victorian era. His works enjoyed unprecedented popularity during his lifetime and, by the 20th century, critics and scholars had recognised him as a literary genius. His novels and short stories are widely read today. Born in Portsmouth, Dickens left school at the age of 12 to work in a boot-blacking factory when his father was incarcerated in a debtors' prison. After three years he returned to school, before he began his literary career as a journalist. Dickens edited a weekly journal for 20 years, wrote 15 novels, five novellas, hundreds of short stories and non-fiction articles, lectured and performed readings extensively, was an indefatigable letter writer, and campaigned vigorously for children's rights, for education, and for other social reforms. Dickens's literary success began with the 1836 serial publication of The Pickwick Papers, a publishing phenomenon—thanks largely to the introduction of the character Sam Weller in the fourth episode—that sparked Pickwick merchandise and spin-offs. Within a few years Dickens had become an international literary celebrity, famous for his humour, satire and keen observation of character and society. His novels, most of them published in monthly or weekly installments, pioneered the serial publication of narrative fiction, which became the dominant Victorian mode for novel publication. Cliffhanger endings in his serial publications kept readers in suspense. The instalment format allowed Dickens to evaluate his audience's reaction, and he often modified his plot and character development based on such feedback. For example, when his wife's chiropodist expressed distress at the way Miss Mowcher in David Copperfield seemed to reflect her own disabilities, Dickens improved the character with positive features. His plots were carefully constructed and he often wove elements from topical events into his narratives. Masses of the illiterate poor would individually pay a halfpenny to have each new monthly episode read to them, opening up and inspiring a new class of readers. His 1843 novella A Christmas Carol remains especially popular and continues to inspire adaptations in every artistic genre. Oliver Twist and Great Expectations are also frequently adapted and, like many of his novels, evoke images of early Victorian London. His 1859 novel A Tale of Two Cities (set in London and Paris) is his best-known work of historical fiction. The most famous celebrity of his era, he undertook, in response to public demand, a series of public reading tours in the later part of his career. The term Dickensian is used to describe something that is reminiscent of Dickens and his writings, such as poor social or working conditions, or comically repulsive characters. Early life Charles Dickens was born on 7 February 1812 at 1 Mile End Terrace (now 393 Commercial Road), Landport in Portsea Island (Portsmouth), Hampshire, the second of eight children of Elizabeth Dickens (née Barrow; 1789–1863) and John Dickens (1785–1851). His father was a clerk in the Navy Pay Office and was temporarily stationed in the district. He asked Christopher Huffam, rigger to His Majesty's Navy, gentleman, and head of an established firm, to act as godfather to Charles. Huffam is thought to be the inspiration for Paul Dombey, the owner of a shipping company in Dickens's novel Dombey and Son (1848). In January 1815, John Dickens was called back to London and the family moved to Norfolk Street, Fitzrovia. When Charles was four, they relocated to Sheerness and thence to Chatham, Kent, where he spent his formative years until the age of 11. His early life seems to have been idyllic, though he thought himself a "very small and not-over-particularly-taken-care-of boy". Charles spent time outdoors, but also read voraciously, including the picaresque novels of Tobias Smollett and Henry Fielding, as well as Robinson Crusoe and Gil Blas. He read and reread The Arabian Nights and the Collected Farces of Elizabeth Inchbald. At age 7 he first saw Joseph Grimaldi—the father of modern clowning—perform at the Star Theatre, Rochester. He later imitated Grimaldi's clowning on several occasions, and would also edit the Memoirs of Joseph Grimaldi. He retained poignant memories of childhood, helped by an excellent memory of people and events, which he used in his writing. His father's brief work as a clerk in the Navy Pay Office afforded him a few years of private education, first at a dame school and then at a school run by William Giles, a dissenter, in Chatham. This period came to an end in June 1822, when John Dickens was recalled to Navy Pay Office headquarters at Somerset House and the family (except for Charles, who stayed behind to finish his final term at school) moved to Camden Town in London. The family had left Kent amidst rapidly mounting debts and, living beyond his means, John Dickens was forced by his creditors into the Marshalsea debtors' prison in Southwark, London in 1824. His wife and youngest children joined him there, as was the practice at the time. Charles, then 12 years old, boarded with Elizabeth Roylance, a family friend, at 112 College Place, Camden Town. Mrs Roylance was "a reduced impoverished old lady, long known to our family", whom Dickens later immortalised, "with a few alterations and embellishments", as "Mrs Pipchin" in Dombey and Son. Later, he lived in a back-attic in the house of an agent for the Insolvent Court, Archibald Russell, "a fat, good-natured, kind old gentleman ... with a quiet old wife" and lame son, in Lant Street in Southwark. They provided the inspiration for the Garlands in The Old Curiosity Shop. On Sundays – with his sister Frances, free from her studies at the Royal Academy of Music – he spent the day at the Marshalsea. Dickens later used the prison as a setting in Little Dorrit. To pay for his board and to help his family, Dickens was forced to leave school and work ten-hour days at Warren's Blacking Warehouse, on Hungerford Stairs, near the present Charing Cross railway station, where he earned six shillings a week pasting labels on pots of boot blacking. The strenuous and often harsh working conditions made a lasting impression on Dickens and later influenced his fiction and essays, becoming the foundation of his interest in the reform of socio-economic and labour conditions, the rigours of which he believed were unfairly borne by the poor. He later wrote that he wondered "how I could have been so easily cast away at such an age". As he recalled to John Forster (from Life of Charles Dickens): When the warehouse was moved to Chandos Street in the smart, busy district of Covent Garden, the boys worked in a room in which the window gave onto the street. Small audiences gathered and watched them at work – in Dickens's biographer Simon Callow's estimation, the public display was "a new refinement added to his misery". A few months after his imprisonment, John Dickens's mother, Elizabeth Dickens, died and bequeathed him £450. On the expectation of this legacy, Dickens was released from prison. Under the Insolvent Debtors Act, Dickens arranged for payment of his creditors and he and his family left the Marshalsea, for the home of Mrs Roylance. Charles's mother, Elizabeth Dickens, did not immediately support his removal from the boot-blacking warehouse. This influenced Dickens's view that a father should rule the family and a mother find her proper sphere inside the home: "I never afterwards forgot, I never shall forget, I never can forget, that my mother was warm for my being sent back." His mother's failure to request his return was a factor in his dissatisfied attitude towards women. Righteous indignation stemming from his own situation and the conditions under which working-class people lived became major themes of his works, and it was this unhappy period in his youth to which he alluded in his favourite, and most autobiographical, novel, David Copperfield: "I had no advice, no counsel, no encouragement, no consolation, no assistance, no support, of any kind, from anyone, that I can call to mind, as I hope to go to heaven!" Dickens was eventually sent to the Wellington House Academy in Camden Town, where he remained until March 1827, having spent about two years there. He did not consider it to be a good school: "Much of the haphazard, desultory teaching, poor discipline punctuated by the headmaster's sadistic brutality, the seedy ushers and general run-down atmosphere, are embodied in Mr Creakle's Establishment in David Copperfield." Dickens worked at the law office of Ellis and Blackmore, attorneys, of Holborn Court, Gray's Inn, as a junior clerk from May 1827 to November 1828. He was a gifted mimic and impersonated those around him: clients, lawyers and clerks. He went to theatres obsessively: he claimed that for at least three years he went to the theatre every day. His favourite actor was Charles Mathews and Dickens learnt his "monopolylogues" (farces in which Mathews played every character) by heart. Then, having learned Gurney's system of shorthand in his spare time, he left to become a freelance reporter. A distant relative, Thomas Charlton, was a freelance reporter at Doctors' Commons and Dickens was able to share his box there to report the legal proceedings for nearly four years. This education was to inform works such as Nicholas Nickleby, Dombey and Son and especially Bleak House, whose vivid portrayal of the machinations and bureaucracy of the legal system did much to enlighten the general public and served as a vehicle for dissemination of Dickens's own views regarding, particularly, the heavy burden on the poor who were forced by circumstances to "go to law". In 1830, Dickens met his first love, Maria Beadnell, thought to have been the model for the character Dora in David Copperfield. Maria's parents disapproved of the courtship and ended the relationship by sending her to school in Paris. Career Journalism and writing In 1832, at the age of 20, Dickens was energetic and increasingly self-confident. He enjoyed mimicry and popular entertainment, lacked a clear, specific sense of what he wanted to become, and yet knew he wanted fame. Drawn to the theatre – he became an early member of the Garrick Club – he landed an acting audition at Covent Garden, where the manager George Bartley and the actor Charles Kemble were to see him. Dickens prepared meticulously and decided to imitate the comedian Charles Mathews, but ultimately he missed the audition because of a cold. Before another opportunity arose, he had set out on his career as a writer. In 1833, Dickens submitted his first story, "A Dinner at Poplar Walk", to the London periodical Monthly Magazine. William Barrow, Dickens's uncle on his mother's side, offered him a job on The Mirror of Parliament and he worked in the House of Commons for the first time early in 1832. He rented rooms at Furnival's Inn and worked as a political journalist, reporting on Parliamentary debates, and he travelled across Britain to cover election campaigns for the Morning Chronicle. His journalism, in the form of sketches in periodicals, formed his first collection of pieces, published in 1836: Sketches by Boz – Boz being a family nickname he employed as a pseudonym for some years. Dickens apparently adopted it from the nickname 'Moses', which he had given to his youngest brother Augustus Dickens, after a character in Oliver Goldsmith's The Vicar of Wakefield. When pronounced by anyone with a head cold, "Moses" became "Boses" – later shortened to Boz. Dickens's own name was considered "queer" by a contemporary critic, who wrote in 1849: "Mr Dickens, as if in revenge for his own queer name, does bestow still queerer ones upon his fictitious creations." Dickens contributed to and edited journals throughout his literary career. In January 1835, the Morning Chronicle launched an evening edition, under the editorship of the Chronicles music critic, George Hogarth. Hogarth invited him to contribute Street Sketches and Dickens became a regular visitor to his Fulham house – excited by Hogarth's friendship with Walter Scott (whom Dickens greatly admired) and enjoying the company of Hogarth's three daughters: Georgina, Mary and 19-year-old Catherine. Dickens made rapid progress both professionally and socially. He began a friendship with William Harrison Ainsworth, the author of the highwayman novel Rookwood (1834), whose bachelor salon in Harrow Road had become the meeting place for a set that included Daniel Maclise, Benjamin Disraeli, Edward Bulwer-Lytton and George Cruikshank. All these became his friends and collaborators, with the exception of Disraeli, and he met his first publisher, John Macrone, at the house. The success of Sketches by Boz led to a proposal from publishers Chapman and Hall for Dickens to supply text to match Robert Seymour's engraved illustrations in a monthly letterpress. Seymour committed suicide after the second instalment and Dickens, who wanted to write a connected series of sketches, hired "Phiz" to provide the engravings (which were reduced from four to two per instalment) for the story. The resulting story became The Pickwick Papers and, although the first few episodes were not successful, the introduction of the Cockney character Sam Weller in the fourth episode (the first to be illustrated by Phiz) marked a sharp climb in its popularity. The final instalment sold 40,000 copies. On the impact of the character, The Paris Review stated, "arguably the most historic bump in English publishing is the Sam Weller Bump." A publishing phenomenon, John Sutherland called The Pickwick Papers "[t]he most important single novel of the Victorian era". The unprecedented success led to numerous spin-offs and merchandise including Pickwick cigars, playing cards, china figurines, Sam Weller puzzles, Weller boot polish and joke books. On the creation of modern mass culture, Nicholas Dames in The Atlantic writes, "Literature" is not a big enough category for Pickwick. It defined its own, a new one that we have learned to call "entertainment." In November 1836, Dickens accepted the position of editor of Bentley's Miscellany, a position he held for three years, until he fell out with the owner. In 1836, as he finished the last instalments of The Pickwick Papers, he began writing the beginning instalments of Oliver Twist – writing as many as 90 pages a month – while continuing work on Bentley's and also writing four plays, the production of which he oversaw. Oliver Twist, published in 1838, became one of Dickens's better known stories and was the first Victorian novel with a child protagonist. On 2 April 1836, after a one-year engagement, and between episodes two and three of The Pickwick Papers, Dickens married Catherine Thomson Hogarth (1815–1879), the daughter of George Hogarth, editor of the Evening Chronicle. They were married in St Luke's Church, Chelsea, London. After a brief honeymoon in Chalk in Kent, the couple returned to lodgings at Furnival's Inn. The first of their ten children, Charles, was born in January 1837 and a few months later the family set up home in Bloomsbury at 48 Doughty Street, London (on which Charles had a three-year lease at £80 a year) from 25 March 1837 until December 1839. Dickens's younger brother Frederick and Catherine's 17-year-old sister Mary Hogarth moved in with them. Dickens became very attached to Mary, and she died in his arms after a brief illness in 1837. Unusually for Dickens, as a consequence of his shock, he stopped working, and he and Catherine stayed at a little farm on Hampstead Heath for a fortnight. Dickens idealised Mary; the character he fashioned after her, Rose Maylie, he found he could not now kill, as he had planned, in his fiction, and, according to Ackroyd, he drew on memories of her for his later descriptions of Little Nell and Florence Dombey. His grief was so great that he was unable to meet the deadline for the June instalment of The Pickwick Papers and had to cancel the Oliver Twist instalment that month as well. The time in Hampstead was the occasion for a growing bond between Dickens and John Forster to develop; Forster soon became his unofficial business manager and the first to read his work. His success as a novelist continued. The young Queen Victoria read both Oliver Twist and The Pickwick Papers, staying up until midnight to discuss them. Nicholas Nickleby (1838–39), The Old Curiosity Shop (1840–41) and, finally, his first historical novel, Barnaby Rudge: A Tale of the Riots of 'Eighty, as part of the Master Humphrey's Clock series (1840–41), were all published in monthly instalments before being made into books. In the midst of all his activity during this period, there was discontent with his publishers and John Macrone was bought off, while Richard Bentley signed over all his rights in Oliver Twist. Other signs of a certain restlessness and discontent emerged; in Broadstairs he flirted with Eleanor Picken, the young fiancée of his solicitor's best friend and one night grabbed her and ran with her down to the sea. He declared they were both to drown there in the "sad sea waves". She finally got free, and afterwards kept her distance. In June 1841, he precipitously set out on a two-month tour of Scotland and then, in September 1841, telegraphed Forster that he had decided to go to America. Master Humphrey's Clock was shut down, though Dickens was still keen on the idea of the weekly magazine, a form he liked, an appreciation that had begun with his childhood reading of the 18th-century magazines Tatler and The Spectator. Dickens was perturbed by the return to power of the Tories, whom he described as "people whom, politically, I despise and abhor." He had been tempted to stand for the Liberals in Reading, but decided against it due to financial straits. He wrote three anti-Tory verse satires ("The Fine Old English Gentleman", "The Quack Doctor's Proclamation", and "Subjects for Painters") which were published in The Examiner. First visit to the United States On 22 January 1842, Dickens and his wife arrived in Boston, Massachusetts, aboard the RMS Britannia during their first trip to the United States and Canada. At this time Georgina Hogarth, another sister of Catherine, joined the Dickens household, now living at Devonshire Terrace, Marylebone to care for the young family they had left behind. She remained with them as housekeeper, organiser, adviser and friend until Dickens's death in 1870. Dickens modelled the character of Agnes Wickfield after Georgina and Mary. He described his impressions in a travelogue, American Notes for General Circulation. In Notes, Dickens includes a powerful condemnation of slavery which he had attacked as early as The Pickwick Papers, correlating the emancipation of the poor in England with the abolition of slavery abroad citing newspaper accounts of runaway slaves disfigured by their masters. In spite of the abolitionist sentiments gleaned from his trip to America, some modern commentators have pointed out inconsistencies in Dickens's views on racial inequality. For instance, he has been criticised for his subsequent acquiescence in Governor Eyre's harsh crackdown during the 1860s Morant Bay rebellion in Jamaica and his failure to join other British progressives in condemning it. From Richmond, Virginia, Dickens returned to Washington, D.C., and started a trek westward, with brief pauses in Cincinnati and Louisville, to St. Louis, Missouri. While there, he expressed a desire to see an American prairie before returning east. A group of 13 men then set out with Dickens to visit Looking Glass Prairie, a trip 30 miles into Illinois. During his American visit, Dickens spent a month in New York City, giving lectures, raising the question of international copyright laws and the pirating of his work in America. He persuaded a group of 25 writers, headed by Washington Irving, to sign a petition for him to take to Congress, but the press were generally hostile to this, saying that he should be grateful for his popularity and that it was mercenary to complain about his work being pirated. The popularity he gained caused a shift in his self-perception according to critic Kate Flint, who writes that he "found himself a cultural commodity, and its circulation had passed out his control", causing him to become interested in and delve into themes of public and personal personas in the next novels. She writes that he assumed a role of "influential commentator", publicly and in his fiction, evident in his next few books. His trip to the U.S. ended with a trip to Canada – Niagara Falls, Toronto, Kingston and Montreal – where he appeared on stage in light comedies. Return to England Soon after his return to England, Dickens began work on the first of his Christmas stories, A Christmas Carol, written in 1843, which was followed by The Chimes in 1844 and The Cricket on the Hearth in 1845. Of these, A Christmas Carol was most popular and, tapping into an old tradition, did much to promote a renewed enthusiasm for the joys of Christmas in Britain and America. The seeds for the story became planted in Dickens's mind during a trip to Manchester to witness the conditions of the manufacturing workers there. This, along with scenes he had recently witnessed at the Field Lane Ragged School, caused Dickens to resolve to "strike a sledge hammer blow" for the poor. As the idea for the story took shape and the writing began in earnest, Dickens became engrossed in the book. He later wrote that as the tale unfolded he "wept and laughed, and wept again" as he "walked about the black streets of London fifteen or twenty miles many a night when all sober folks had gone to bed". After living briefly in Italy (1844), Dickens travelled to Switzerland (1846), where he began work on Dombey and Son (1846–48). This and David Copperfield (1849–50) mark a significant artistic break in Dickens's career as his novels became more serious in theme and more carefully planned than his early works. At about this time, he was made aware of a large embezzlement at the firm where his brother, Augustus, worked (John Chapman & Co). It had been carried out by Thomas Powell, a clerk, who was on friendly terms with Dickens and who had acted as mentor to Augustus when he started work. Powell was also an author and poet and knew many of the famous writers of the day. After further fraudulent activities, Powell fled to New York and published a book called The Living Authors of England with a chapter on Charles Dickens, who was not amused by what Powell had written. One item that seemed to have annoyed him was the assertion that he had based the character of Paul Dombey (Dombey and Son) on Thomas Chapman, one of the principal partners at John Chapman & Co. Dickens immediately sent a letter to Lewis Gaylord Clark, editor of the New York literary magazine The Knickerbocker, saying that Powell was a forger and thief. Clark published the letter in the New-York Tribune and several other papers picked up on the story. Powell began proceedings to sue these publications and Clark was arrested. Dickens, realising that he had acted in haste, contacted John Chapman & Co to seek written confirmation of Powell's guilt. Dickens did receive a reply confirming Powell's embezzlement, but once the directors realised this information might have to be produced in court, they refused to make further disclosures. Owing to the difficulties of providing evidence in America to support his accusations, Dickens eventually made a private settlement with Powell out of court. Philanthropy Angela Burdett Coutts, heir to the Coutts banking fortune, approached Dickens in May 1846 about setting up a home for the redemption of fallen women of the working class. Coutts envisioned a home that would replace the punitive regimes of existing institutions with a reformative environment conducive to education and proficiency in domestic household chores. After initially resisting, Dickens eventually founded the home, named Urania Cottage, in the Lime Grove area of Shepherd's Bush, which he managed for ten years, setting the house rules, reviewing the accounts and interviewing prospective residents. Emigration and marriage were central to Dickens's agenda for the women on leaving Urania Cottage, from which it is estimated that about 100 women graduated between 1847 and 1859. Religious views As a young man, Dickens expressed a distaste for certain aspects of organised religion. In 1836, in a pamphlet titled Sunday Under Three Heads, he defended the people's right to pleasure, opposing a plan to prohibit games on Sundays. "Look into your churches – diminished congregations and scanty attendance. People have grown sullen and obstinate, and are becoming disgusted with the faith which condemns them to such a day as this, once in every seven. They display their feeling by staying away [from church]. Turn into the streets [on a Sunday] and mark the rigid gloom that reigns over everything around." Dickens honoured the figure of Jesus Christ. He is regarded as a professing Christian. His son, Henry Fielding Dickens, described him as someone who "possessed deep religious convictions". In the early 1840s, he had shown an interest in Unitarian Christianity and Robert Browning remarked that "Mr Dickens is an enlightened Unitarian." Professor Gary Colledge has written that he "never strayed from his attachment to popular lay Anglicanism". Dickens authored a work called The Life of Our Lord (1846), a book about the life of Christ, written with the purpose of sharing his faith with his children and family. In a scene from David Copperfield, Dickens echoed Geoffrey Chaucer's use of Luke 23:34 from Troilus and Criseyde (Dickens held a copy in his library), with G. K. Chesterton writing, "among the great canonical English authors, Chaucer and Dickens have the most in common." Dickens disapproved of Roman Catholicism and 19th-century evangelicalism, seeing both as extremes of Christianity and likely to limit personal expression, and was critical of what he saw as the hypocrisy of religious institutions and philosophies like spiritualism, all of which he considered deviations from the true spirit of Christianity, as shown in the book he wrote for his family in 1846. While Dickens advocated equal rights for Catholics in England, he strongly disliked how individual civil liberties were often threatened in countries where Catholicism predominated and referred to the Catholic Church as "that curse upon the world." Dickens also rejected the Evangelical conviction that the Bible was the infallible word of God. His ideas on Biblical interpretation were similar to the Liberal Anglican Arthur Penrhyn Stanley's doctrine of "progressive revelation." Leo Tolstoy and Fyodor Dostoyevsky referred to Dickens as "that great Christian writer". Middle years In December 1845, Dickens took up the editorship of the London-based Daily News, a liberal paper through which Dickens hoped to advocate, in his own words, "the Principles of Progress and Improvement, of Education and Civil and Religious Liberty and Equal Legislation." Among the other contributors Dickens chose to write for the paper were the radical economist Thomas Hodgskin and the social reformer Douglas William Jerrold, who frequently attacked the Corn Laws. Dickens lasted only ten weeks on the job before resigning due to a combination of exhaustion and frustration with one of the paper's co-owners. A Francophile, Dickens often holidayed in France and, in a speech delivered in Paris in 1846 in French, called the French "the first people in the universe". During his visit to Paris, Dickens met the French literati Alexandre Dumas, Victor Hugo, Eugène Scribe, Théophile Gautier, François-René de Chateaubriand and Eugène Sue. In early 1849, Dickens started to write David Copperfield. It was published between 1849 and 1850. In Dickens's biography, Life of Charles Dickens (1872), John Forster wrote of David Copperfield, "underneath the fiction lay something of the author's life". It was Dickens's personal favourite among his own novels, as he wrote in the author's preface to the 1867 edition of the novel. In late November 1851, Dickens moved into Tavistock House where he wrote Bleak House (1852–53), Hard Times (1854) and Little Dorrit (1856). It was here that he indulged in the amateur theatricals described in Forster's Life of Charles Dickens. During this period, he worked closely with the novelist and playwright Wilkie Collins. In 1856, his income from writing allowed him to buy Gads Hill Place in Higham, Kent. As a child, Dickens had walked past the house and dreamed of living in it. The area was also the scene of some of the events of Shakespeare's Henry IV, Part 1 and this literary connection pleased him. During this time Dickens was also the publisher, editor and a major contributor to the journals Household Words (1850–1859) and All the Year Round (1858–1870). In 1854, at the behest of Sir John Franklin's widow Lady Jane, Dickens viciously attacked Arctic explorer John Rae in Household Words for his report to the Admiralty, based on interviews with local Inuit, that the members of Franklin's lost expedition had resorted to cannibalism. These attacks would later be expanded on his 1856 play The Frozen Deep, which satirizes Rae and the Inuit. 20th century archaeology work in King William Island later confirmed that the members of the Franklin expedition resorted to cannibalism. In 1855, when Dickens's good friend and Liberal MP Austen Henry Layard formed an Administrative Reform Association to demand significant reforms of Parliament, Dickens joined and volunteered his resources in support of Layard's cause. With the exception of Lord John Russell, who was the only leading politician in whom Dickens had any faith and to whom he later dedicated A Tale of Two Cities, Dickens believed that the political aristocracy and their incompetence were the death of England. When he and Layard were accused of fomenting class conflict, Dickens replied that the classes were already in opposition and the fault was with the aristocratic class. Dickens used his pulpit in Household Words to champion the Reform Association. He also commented on foreign affairs, declaring his support for Giuseppe Garibaldi and Giuseppe Mazzini, helping raise funds for their campaigns and stating that "a united Italy would be of vast importance to the peace of the world, and would be a rock in Louis Napoleon's way," and that "I feel for Italy almost as if I were an Italian born." Following the Indian Mutiny of 1857, Dickens joined in the widespread criticism of the East India Company for its role in the event, but reserved his fury for Indians, wishing that he was the commander-in-chief in India so that he would be able to, "do my utmost to exterminate the Race upon whom the stain of the late cruelties rested." In 1857, Dickens hired professional actresses for The Frozen Deep, written by him and his protégé, Wilkie Collins. Dickens fell in love with one of the actresses, Ellen Ternan, and this passion was to last the rest of his life. In 1858, when Dickens was 45 and Ternan 18, divorce was nearly unthinkable for someone as famous as he was. After publicly accusing Catherine of not loving their children and suffering from "a mental disorder", statements that disgusted his contemporaries, including Elizabeth Barrett Browning, Dickens attempted to have Catherine institutionalized. When this scheme failed, they separated. When Catherine left, never to see her husband again, she took with her one child, leaving the other children to be raised by her sister Georgina, who chose to stay at Gads Hill. During this period, whilst pondering a project to give public readings for his own profit, Dickens was approached through a charitable appeal by Great Ormond Street Hospital to help it survive its first major financial crisis. His "Drooping Buds" essay in Household Words earlier on 3 April 1852 was considered by the hospital's founders to have been the catalyst for the hospital's success. Dickens, whose philanthropy was well-known, was asked by his friend, the hospital's founder Charles West, to preside over the appeal, and he threw himself into the task, heart and soul. Dickens's public readings secured sufficient funds for an endowment to put the hospital on a sound financial footing; one reading on 9 February 1858 alone raised £3,000. After separating from Catherine, Dickens undertook a series of hugely popular and remunerative reading tours which, together with his journalism, were to absorb most of his creative energies for the next decade, in which he was to write only two more novels. His first reading tour, lasting from April 1858 to February 1859, consisted of 129 appearances in 49 towns throughout England, Scotland and Ireland. Dickens's continued fascination with the theatrical world was written into the theatre scenes in Nicholas Nickleby, but more importantly he found an outlet in public readings. In 1866, he undertook a series of public readings in England and Scotland, with more the following year in England and Ireland. Other works soon followed, including A Tale of Two Cities (1859) and Great Expectations (1861), which were resounding successes. Set in London and Paris, A Tale of Two Cities is his best-known work of historical fiction and includes the famous opening sentence which begins with "It was the best of times, it was the worst of times." It is regularly cited as one of the best-selling novels of all time. Themes in Great Expectations include wealth and poverty, love and rejection, and the eventual triumph of good over evil. In early September 1860, in a field behind Gads Hill, Dickens made a bonfire of most of his correspondence; only those letters on business matters were spared. Since Ellen Ternan also destroyed all of his letters to her, the extent of the affair between the two remains speculative. In the 1930s, Thomas Wright recounted that Ternan had unburdened herself to a Canon Benham and gave currency to rumours they had been lovers. That the two had a son who died in infancy was alleged by Dickens's daughter, Kate Perugini, whom Gladys Storey had interviewed before her death in 1929. Storey published her account in Dickens and Daughter, but no contemporary evidence exists. On his death, Dickens settled an annuity on Ternan which made her financially independent. Claire Tomalin's book, The Invisible Woman, argues that Ternan lived with Dickens secretly for the last 13 years of his life. The book was subsequently turned into a play, Little Nell, by Simon Gray, and a 2013 film. In the same period, Dickens furthered his interest in the paranormal, becoming one of the early members of The Ghost Club. In June 1862, he was offered £10,000 for a reading tour of Australia. He was enthusiastic, and even planned a travel book, The Uncommercial Traveller Upside Down, but ultimately decided against the tour. Two of his sons, Alfred D'Orsay Tennyson Dickens and Edward Bulwer Lytton Dickens, migrated to Australia, Edward becoming a member of the Parliament of New South Wales as Member for Wilcannia between 1889 and 1894. Later life On 9 June 1865, while returning from Paris with Ellen Ternan, Dickens was involved in the Staplehurst rail crash in Kent. The train's first seven carriages plunged off a cast iron bridge that was under repair. The only first-class carriage to remain on the track was the one in which Dickens was travelling. Before rescuers arrived, Dickens tended and comforted the wounded and the dying with a flask of brandy and a hat refreshed with water, and saved some lives. Before leaving, he remembered the unfinished manuscript for Our Mutual Friend, and he returned to his carriage to retrieve it. Dickens later used the experience of the crash as material for his short ghost story, "The Signal-Man", in which the central character has a premonition of his own death in a rail crash. He also based the story on several previous rail accidents, such as the Clayton Tunnel rail crash in Sussex of 1861. Dickens managed to avoid an appearance at the inquest to avoid disclosing that he had been travelling with Ternan and her mother, which would have caused a scandal. After the crash, Dickens was nervous when travelling by train and would use alternative means when available. In 1868 he wrote, "I have sudden vague rushes of terror, even when riding in a hansom cab, which are perfectly unreasonable but quite insurmountable." Dickens's son, Henry, recalled, "I have seen him sometimes in a railway carriage when there was a slight jolt. When this happened he was almost in a state of panic and gripped the seat with both hands." Second visit to the United States While he contemplated a second visit to the United States, the outbreak of the Civil War in America in 1861 delayed his plans. On 9 November 1867, over two years after the war, Dickens set sail from Liverpool for his second American reading tour. Landing in Boston, he devoted the rest of the month to a round of dinners with such notables as Ralph Waldo Emerson, Henry Wadsworth Longfellow and his American publisher, James T. Fields. In early December, the readings began. He performed 76 readings, netting £19,000, from December 1867 to April 1868. Dickens shuttled between Boston and New York, where he gave 22 readings at Steinway Hall. Although he had started to suffer from what he called the "true American catarrh", he kept to a schedule that would have challenged a much younger man, even managing to squeeze in some sleighing in Central Park. During his travels, he saw a change in the people and the circumstances of America. His final appearance was at a banquet the American Press held in his honour at Delmonico's on 18 April, when he promised never to denounce America again. By the end of the tour Dickens could hardly manage solid food, subsisting on champagne and eggs beaten in sherry. On 23 April he boarded the Cunard liner to return to Britain, barely escaping a federal tax lien against the proceeds of his lecture tour. Farewell readings In 1868–69, Dickens gave a series of "farewell readings" in England, Scotland and Ireland, beginning on 6 October. He managed, of a contracted 100 readings, to give 75 in the provinces, with a further 12 in London. As he pressed on he was affected by giddiness and fits of paralysis. He had a stroke on 18 April 1869 in Chester. He collapsed on 22 April 1869, at Preston, Lancashire; on doctor's advice, the tour was cancelled. After further provincial readings were cancelled, he began work on his final novel, The Mystery of Edwin Drood. It was fashionable in the 1860s to 'do the slums' and, in company, Dickens visited opium dens in Shadwell, where he witnessed an elderly addict called "Laskar Sal", who formed the model for "Opium Sal" in Edwin Drood. After Dickens regained enough strength, he arranged, with medical approval, for a final series of readings to partly make up to his sponsors what they had lost due to his illness. There were 12 performances, on 11 January to 15 March 1870; the last at 8:00pm at St. James's Hall, London. Though in grave health by then, he read A Christmas Carol and The Trial from Pickwick. On 2 May, he made his last public appearance at a Royal Academy banquet in the presence of the Prince and Princess of Wales, paying a special tribute on the death of his friend, illustrator Daniel Maclise. Death On 8 June 1870, Dickens had another stroke at his home after a full day's work on Edwin Drood. He never regained consciousness and, the next day, he died at Gads Hill Place. Biographer Claire Tomalin has suggested Dickens was actually in Peckham when he had had the stroke and his mistress Ellen Ternan and her maids had him taken back to Gads Hill so that the public would not know the truth about their relationship. Contrary to his wish to be buried at Rochester Cathedral "in an inexpensive, unostentatious, and strictly private manner", he was laid to rest in the Poets' Corner of Westminster Abbey. A printed epitaph circulated at the time of the funeral reads: A letter from Dickens to the Clerk of the Privy Council in March indicates he'd been offered and had accepted a baronetcy, which was not gazetted before his death. His last words were "On the ground", in response to his sister-in-law Georgina's request that he lie down. On Sunday, 19 June 1870, five days after Dickens was buried in the Abbey, Dean Arthur Penrhyn Stanley delivered a memorial elegy, lauding "the genial and loving humorist whom we now mourn", for showing by his own example "that even in dealing with the darkest scenes and the most degraded characters, genius could still be clean, and mirth could be innocent". Pointing to the fresh flowers that adorned the novelist's grave, Stanley assured those present that "the spot would thenceforth be a sacred one with both the New World and the Old, as that of the representative of literature, not of this island only, but of all who speak our English tongue." In his will, drafted more than a year before his death, Dickens left the care of his £80,000 estate (£ in ) to his long-time colleague John Forster and his "best and truest friend" Georgina Hogarth who, along with Dickens's two sons, also received a tax-free sum of £8,000 (equivalent to £ in ). Although Dickens and his wife had been separated for several years at the time of his death, he provided her with an annual income of £600 (£ in ) and made her similar allowances in his will. He also bequeathed £19 19s (£ in ) to each servant in his employment at the time of his death. Literary style Dickens's approach to the novel is influenced by various things, including the picaresque novel tradition, melodrama and the novel of sensibility. According to Ackroyd, other than these, perhaps the most important literary influence on him was derived from the fables of The Arabian Nights. Satire and irony are central to the picaresque novel. Comedy is also an aspect of the British picaresque novel tradition of Laurence Sterne, Henry Fielding and Tobias Smollett. Fielding's Tom Jones was a major influence on the 19th-century novelist including Dickens, who read it in his youth and named a son Henry Fielding Dickens after him. Influenced by Gothic fiction—a literary genre that began with The Castle of Otranto (1764) by Horace Walpole—Dickens incorporated Gothic imagery, settings and plot devices in his works. Victorian gothic moved from castles and abbeys into contemporary urban environments: in particular London, such as Dickens's Oliver Twist and Bleak House. The jilted bride Miss Havisham from Great Expectations is one of Dickens's best-known gothic creations; living in a ruined mansion, her bridal gown effectively doubles as her funeral shroud. No other writer had such a profound influence on Dickens as William Shakespeare. On Dickens's veneration of Shakespeare, Alfred Harbage wrote "No one is better qualified to recognise literary genius than a literary genius"— A Kind of Power: The Shakespeare-Dickens Analogy (1975). Regarding Shakespeare as "the great master" whose plays "were an unspeakable source of delight", Dickens's lifelong affinity with the playwright included seeing theatrical productions of his plays in London and putting on amateur dramatics with friends in his early years. In 1838 Dickens travelled to Stratford-upon-Avon and visited the house in which Shakespeare was born, leaving his autograph in the visitors' book. Dickens would draw on this experience in his next work, Nicholas Nickleby (1838–39), expressing the strength of feeling experienced by visitors to Shakespeare's birthplace: the character Mrs Wititterly states, "I don't know how it is, but after you've seen the place and written your name in the little book, somehow or other you seem to be inspired; it kindles up quite a fire within one." Dickens's writing style is marked by a profuse linguistic creativity. Satire, flourishing in his gift for caricature, is his forte. An early reviewer compared him to Hogarth for his keen practical sense of the ludicrous side of life, though his acclaimed mastery of varieties of class idiom may in fact mirror the conventions of contemporary popular theatre. Dickens worked intensively on developing arresting names for his characters that would reverberate with associations for his readers and assist the development of motifs in the storyline, giving what one critic calls an "allegorical impetus" to the novels' meanings. To cite one of numerous examples, the name Mr Murdstone in David Copperfield conjures up twin allusions to murder and stony coldness. His literary style is also a mixture of fantasy and realism. His satires of British aristocratic snobbery – he calls one character the "Noble Refrigerator" – are often popular. Comparing orphans to stocks and shares, people to tug boats or dinner-party guests to furniture are just some of Dickens's acclaimed flights of fancy. The author worked closely with his illustrators, supplying them with a summary of the work at the outset and thus ensuring that his characters and settings were exactly how he envisioned them. He briefed the illustrator on plans for each month's instalment so that work could begin before he wrote them. Marcus Stone, illustrator of Our Mutual Friend, recalled that the author was always "ready to describe down to the minutest details the personal characteristics, and ... life-history of the creations of his fancy". Dickens employs Cockney English in many of his works, denoting working-class Londoners. Cockney grammar appears in terms such as ain't, and consonants in words are frequently omitted, as in 'ere (here) and wot (what). An example of this usage is in Oliver Twist. The Artful Dodger uses cockney slang which is juxtaposed with Oliver's 'proper' English, when the Dodger repeats Oliver saying "seven" with "sivin". Characters Dickens's biographer Claire Tomalin regards him as the greatest creator of character in English fiction after Shakespeare. Dickensian characters are amongst the most memorable in English literature, especially so because of their typically whimsical names. The likes of Ebenezer Scrooge, Tiny Tim, Jacob Marley and Bob Cratchit (A Christmas Carol); Oliver Twist, The Artful Dodger, Fagin and Bill Sikes (Oliver Twist); Pip, Miss Havisham and Abel Magwitch (Great Expectations); Sydney Carton, Charles Darnay and Madame Defarge (A Tale of Two Cities); David Copperfield, Uriah Heep and Mr Micawber (David Copperfield); Daniel Quilp and Nell Trent (The Old Curiosity Shop), Samuel Pickwick and Sam Weller (The Pickwick Papers); and Wackford Squeers (Nicholas Nickleby) are so well known as to be part and parcel of popular culture, and in some cases have passed into ordinary language: a scrooge, for example, is a miser or someone who dislikes Christmas festivity. His characters were often so memorable that they took on a life of their own outside his books. "Gamp" became a slang expression for an umbrella from the character Mrs Gamp, and "Pickwickian", "Pecksniffian" and "Gradgrind" all entered dictionaries due to Dickens's original portraits of such characters who were, respectively, quixotic, hypocritical and vapidly factual. The character that made Dickens famous, Sam Weller became known for his Wellerisms—one-liners that turned proverbs on their heads. Many were drawn from real life: Mrs Nickleby is based on his mother, although she did not recognise herself in the portrait, just as Mr Micawber is constructed from aspects of his father's 'rhetorical exuberance'; Harold Skimpole in Bleak House is based on James Henry Leigh Hunt; his wife's dwarfish chiropodist recognised herself in Miss Mowcher in David Copperfield. Perhaps Dickens's impressions on his meeting with Hans Christian Andersen informed the delineation of Uriah Heep (a term synonymous with sycophant). Virginia Woolf maintained that "we remodel our psychological geography when we read Dickens" as he produces "characters who exist not in detail, not accurately or exactly, but abundantly in a cluster of wild yet extraordinarily revealing remarks". T. S. Eliot wrote that Dickens "excelled in character; in the creation of characters of greater intensity than human beings". One "character" vividly drawn throughout his novels is London itself. Dickens described London as a magic lantern, inspiring the places and people in many of his novels. From the coaching inns on the outskirts of the city to the lower reaches of the Thames, all aspects of the capital – Dickens's London – are described over the course of his body of work. Walking the streets (particularly around London) formed an integral part of his writing life, stoking his creativity. Dickens was known to regularly walk at least a dozen miles (19 km) per day, and once wrote, "If I couldn't walk fast and far, I should just explode and perish." Autobiographical elements Authors frequently draw their portraits of characters from people they have known in real life. David Copperfield is regarded by many as a veiled autobiography of Dickens. The scenes of interminable court cases and legal arguments in Bleak House reflect Dickens's experiences as a law clerk and court reporter, and in particular his direct experience of the law's procedural delay during 1844 when he sued publishers in Chancery for breach of copyright. Dickens's father was sent to prison for debt and this became a common theme in many of his books, with the detailed depiction of life in the Marshalsea prison in Little Dorrit resulting from Dickens's own experiences of the institution. Lucy Stroughill, a childhood sweetheart, may have affected several of Dickens's portraits of girls such as Little Em'ly in David Copperfield and Lucie Manette in A Tale of Two Cities. Dickens may have drawn on his childhood experiences, but he was also ashamed of them and would not reveal that this was where he gathered his realistic accounts of squalor. Very few knew the details of his early life until six years after his death, when John Forster published a biography on which Dickens had collaborated. Though Skimpole brutally sends up Leigh Hunt, some critics have detected in his portrait features of Dickens's own character, which he sought to exorcise by self-parody. Episodic writing A pioneer of the serial publication of narrative fiction, Dickens wrote most of his major novels in monthly or weekly instalments in journals such as Master Humphrey's Clock and Household Words, later reprinted in book form. These instalments made the stories affordable and accessible, with the audience more evenly distributed across income levels than previous. His instalment format inspired a narrative that he would explore and develop throughout his career, and the regular cliffhangers made each new episode widely anticipated. When The Old Curiosity Shop was being serialised, American fans waited at the docks in New York harbour, shouting out to the crew of an incoming British ship, "Is little Nell dead?" Dickens's talent was to incorporate this episodic writing style but still end up with a coherent novel at the end. Another important impact of Dickens's episodic writing style resulted from his exposure to the opinions of his readers and friends. His friend Forster had a significant hand in reviewing his drafts, an influence that went beyond matters of punctuation. He toned down melodramatic and sensationalist exaggerations, cut long passages (such as the episode of Quilp's drowning in The Old Curiosity Shop), and made suggestions about plot and character. It was he who suggested that Charley Bates should be redeemed in Oliver Twist. Dickens had not thought of killing Little Nell and it was Forster who advised him to entertain this possibility as necessary to his conception of the heroine. At the helm in popularising cliffhangers and serial publications in Victorian literature, Dickens's influence can also be seen in television soap operas and film series, with The Guardian stating "the DNA of Dickens's busy, episodic storytelling, delivered in instalments and rife with cliffhangers and diversions, is traceable in everything." His serialisation of his novels also drew comments from other writers. In Scottish author Robert Louis Stevenson's novel The Wrecker, Captain Nares, investigating an abandoned ship, remarked: "See! They were writing up the log," said Nares, pointing to the ink-bottle. "Caught napping, as usual. I wonder if there ever was a captain yet that lost a ship with his log-book up to date? He generally has about a month to fill up on a clean break, like Charles Dickens and his serial novels." Social commentary Dickens's novels were, among other things, works of social commentary. Simon Callow states, "From the moment he started to write, he spoke for the people, and the people loved him for it." He was a fierce critic of the poverty and social stratification of Victorian society. In a New York address, he expressed his belief that "Virtue shows quite as well in rags and patches as she does in purple and fine linen". Dickens's second novel, Oliver Twist (1839), shocked readers with its images of poverty and crime: it challenged middle class polemics about criminals, making impossible any pretence to ignorance about what poverty entailed. At a time when Britain was the major economic and political power of the world, Dickens highlighted the life of the forgotten poor and disadvantaged within society. Through his journalism he campaigned on specific issues – such as sanitation and the workhouse – but his fiction probably demonstrated its greatest prowess in changing public opinion in regard to class inequalities. He often depicted the exploitation and oppression of the poor and condemned the public officials and institutions that not only allowed such abuses to exist, but flourished as a result. His most strident indictment of this condition is in Hard Times (1854), Dickens's only novel-length treatment of the industrial working class. In this work, he uses vitriol and satire to illustrate how this marginalised social stratum was termed "Hands" by the factory owners; that is, not really "people" but rather only appendages of the machines they operated. His writings inspired others, in particular journalists and political figures, to address such problems of class oppression. For example, the prison scenes in The Pickwick Papers are claimed to have been influential in having the Fleet Prison shut down. Karl Marx asserted that Dickens "issued to the world more political and social truths than have been uttered by all the professional politicians, publicists and moralists put together". George Bernard Shaw even remarked that Great Expectations was more seditious than Marx's Das Kapital. The exceptional popularity of Dickens's novels, even those with socially oppositional themes (Bleak House, 1853; Little Dorrit, 1857; Our Mutual Friend, 1865), not only underscored his ability to create compelling storylines and unforgettable characters, but also ensured that the Victorian public confronted issues of social justice that had commonly been ignored. It has been argued that his technique of flooding his narratives with an 'unruly superfluity of material' that, in the gradual dénouement, yields up an unsuspected order, influenced the organisation of Charles Darwin's On the Origin of Species. Literary techniques Dickens is often described as using idealised characters and highly sentimental scenes to contrast with his caricatures and the ugly social truths he reveals. The story of Nell Trent in The Old Curiosity Shop (1841) was received as extraordinarily moving by contemporary readers but viewed as ludicrously sentimental by Oscar Wilde. "One must have a heart of stone to read the death of little Nell", he said in a famous remark, "without dissolving into tears ... of laughter." G. K. Chesterton stated, "It is not the death of little Nell, but the life of little Nell, that I object to", arguing that the maudlin effect of his description of her life owed much to the gregarious nature of Dickens's grief, his "despotic" use of people's feelings to move them to tears in works like this. The question as to whether Dickens belongs to the tradition of the sentimental novel is debatable. Valerie Purton, in her book Dickens and the Sentimental Tradition, sees him continuing aspects of this tradition, and argues that his "sentimental scenes and characters [are] as crucial to the overall power of the novels as his darker or comic figures and scenes", and that "Dombey and Son is [ ... ] Dickens's greatest triumph in the sentimentalist tradition". The Encyclopædia Britannica online comments that, despite "patches of emotional excess", such as the reported death of Tiny Tim in A Christmas Carol (1843), "Dickens cannot really be termed a sentimental novelist". In Oliver Twist Dickens provides readers with an idealised portrait of a boy so inherently and unrealistically good that his values are never subverted by either brutal orphanages or coerced involvement in a gang of young pickpockets. While later novels also centre on idealised characters (Esther Summerson in Bleak House and Amy Dorrit in Little Dorrit), this idealism serves only to highlight Dickens's goal of poignant social commentary. Dickens's fiction, reflecting what he believed to be true of his own life, makes frequent use of coincidence, either for comic effect or to emphasise the idea of providence. For example, Oliver Twist turns out to be the lost nephew of the upper-class family that rescues him from the dangers of the pickpocket group. Such coincidences are a staple of 18th-century picaresque novels, such as Henry Fielding's Tom Jones, which Dickens enjoyed reading as a youth. Reputation Dickens was the most popular novelist of his time, and remains one of the best-known and most-read of English authors. His works have never gone out of print, and have been adapted continually for the screen since the invention of cinema, with at least 200 motion pictures and TV adaptations based on Dickens's works documented. Many of his works were adapted for the stage during his own lifetime – early productions included The Haunted Man which was performed in the West End's Adelphi Theatre in 1848 – and, as early as 1901, the British silent film Scrooge, or, Marley's Ghost was made by Walter R. Booth. Contemporaries such as publisher Edward Lloyd cashed in on Dickens's popularity with cheap imitations of his novels, resulting in his own popular 'penny dreadfuls'. From the beginning of his career in the 1830s, Dickens's achievements in English literature were compared to those of Shakespeare. Dickens created some of the world's best-known fictional characters and is regarded by many as the greatest British novelist of the Victorian era. His literary reputation, however began to decline with the publication of Bleak House in 1852–53. Philip Collins calls Bleak House 'a crucial item in the history of Dickens's reputation. Reviewers and literary figures during the 1850s, 1860s and 1870s, saw a "drear decline" in Dickens, from a writer of "bright sunny comedy ... to dark and serious social" commentary. The Spectator called Bleak House "a heavy book to read through at once ... dull and wearisome as a serial"; Richard Simpson, in The Rambler, characterised Hard Times as "this dreary framework"; Fraser's Magazine thought Little Dorrit "decidedly the worst of his novels". All the same, despite these "increasing reservations amongst reviewers and the chattering classes, 'the public never deserted its favourite'". Dickens's popular reputation remained unchanged, sales continued to rise, and Household Words and later All the Year Round were highly successful. As his career progressed, Dickens's fame and the demand for his public readings were unparalleled. In 1868 The Times wrote, "Amid all the variety of 'readings', those of Mr Charles Dickens stand alone." A Dickens biographer, Edgar Johnson, wrote in the 1950s: "It was [always] more than a reading; it was an extraordinary exhibition of acting that seized upon its auditors with a mesmeric possession." Juliet John backed the claim for Dickens "to be called the first self-made global media star of the age of mass culture." Comparing his reception at public readings to those of a contemporary pop star, The Guardian states, "People sometimes fainted at his shows. His performances even saw the rise of that modern phenomenon, the 'speculator' or ticket tout (scalpers) – the ones in New York City escaped detection by borrowing respectable-looking hats from the waiters in nearby restaurants." Among fellow writers, there was a range of opinions on Dickens. Poet laureate, William Wordsworth (1770–1850), thought him a "very talkative, vulgar young person", adding he had not read a line of his work, while novelist George Meredith (1828–1909), found Dickens "intellectually lacking". In 1888 Leslie Stephen commented in the Dictionary of National Biography that "if literary fame could be safely measured by popularity with the half-educated, Dickens must claim the highest position among English novelists". Anthony Trollope's Autobiography famously declared Thackeray, not Dickens, to be the greatest novelist of the age. However, both Leo Tolstoy and Fyodor Dostoyevsky were admirers. Dostoyevsky commented: "We understand Dickens in Russia, I am convinced, almost as well as the English, perhaps even with all the nuances. It may well be that we love him no less than his compatriots do. And yet how original is Dickens, and how very English!" Tolstoy referred to David Copperfield as his favourite book, and he later adopted the novel as "a model for his own autobiographical reflections". French writer Jules Verne called Dickens his favourite writer, writing his novels "stand alone, dwarfing all others by their amazing power and felicity of expression". Dutch painter Vincent van Gogh was inspired by Dickens's novels in several of his paintings like Vincent's Chair and in an 1889 letter to his sister stated that reading Dickens, especially A Christmas Carol, was one of the things that was keeping him from committing suicide. Oscar Wilde generally disparaged his depiction of character, while admiring his gift for caricature. Henry James denied him a premier position, calling him "the greatest of superficial novelists": Dickens failed to endow his characters with psychological depth, and the novels, "loose baggy monsters", betrayed a "cavalier organisation". Joseph Conrad described his own childhood in bleak Dickensian terms, noting he had "an intense and unreasoning affection" for Bleak House dating back to his boyhood. The novel influenced his own gloomy portrait of London in The Secret Agent (1907). Virginia Woolf had a love-hate relationship with Dickens, finding his novels "mesmerizing" while reproving him for his sentimentalism and a commonplace style. Around 1940–41, the attitude of the literary critics began to warm towards Dickens – led by George Orwell in Inside the Whale and Other Essays (March 1940), Edmund Wilson in The Wound and the Bow (1941) and Humphry House in Dickens and his World. However, even in 1948, F. R. Leavis, in The Great Tradition, asserted that "the adult mind doesn't as a rule find in Dickens a challenge to an unusual and sustained seriousness"; Dickens was indeed a great genius, "but the genius was that of a great entertainer", though he later changed his opinion with Dickens the Novelist (1970, with Q. D. (Queenie) Leavis): "Our purpose", they wrote, "is to enforce as unanswerably as possible the conviction that Dickens was one of the greatest of creative writers". In 1944, Soviet film director and film theorist Sergei Eisenstein wrote an essay on Dickens's influence on cinema, such as cross-cutting – where two stories run alongside each other, as seen in novels such as Oliver Twist. In the 1950s, "a substantial reassessment and re-editing of the works began, and critics found his finest artistry and greatest depth to be in the later novels: Bleak House, Little Dorrit, and Great Expectations – and (less unanimously) in Hard Times and Our Mutual Friend". Dickens was a favourite author of Roald Dahl; the best-selling children's author would include three of Dickens's novels among those read by the title character in his 1988 novel Matilda. An avid reader of Dickens, in 2005, Paul McCartney named Nicholas Nickleby his favourite novel. On Dickens he states, "I like the world that he takes me to. I like his words; I like the language", adding, "A lot of my stuff – it's kind of Dickensian." Screenwriter Jonathan Nolan's screenplay for The Dark Knight Rises (2012) was inspired by A Tale of Two Cities, with Nolan calling the depiction of Paris in the novel "one of the most harrowing portraits of a relatable, recognisable civilisation that completely folded to pieces". On 7 February 2012, the 200th anniversary of Dickens's birth, Philip Womack wrote in The Telegraph: "Today there is no escaping Charles Dickens. Not that there has ever been much chance of that before. He has a deep, peculiar hold upon us". Legacy Museums and festivals celebrating Dickens's life and works exist in many places with which Dickens was associated. These include the Charles Dickens Museum in London, the historic home where he wrote Oliver Twist, The Pickwick Papers and Nicholas Nickleby; and the Charles Dickens Birthplace Museum in Portsmouth, the house in which he was born. The original manuscripts of many of his novels, as well as printers' proofs, first editions, and illustrations from the collection of Dickens's friend John Forster are held at the Victoria and Albert Museum. Dickens's will stipulated that no memorial be erected in his honour; nonetheless, a life-size bronze statue of Dickens entitled Dickens and Little Nell, cast in 1890 by Francis Edwin Elwell, stands in Clark Park in the Spruce Hill neighbourhood of Philadelphia, Pennsylvania. Another life-size statue of Dickens is located at Centennial Park in Sydney, Australia. In 1960 a bas-relief sculpture of Dickens, notably featuring characters from his books, was commissioned from sculptor Estcourt J Clack to adorn the office building built on the site of his former home at 1 Devonshire Terrace, London. In 2014, a life-size statue was unveiled near his birthplace in Portsmouth on the 202nd anniversary of his birth; this was supported by his great-great-grandsons, Ian and Gerald Dickens. A Christmas Carol is most probably his best-known story, with frequent new adaptations. It is also the most-filmed of Dickens's stories, with many versions dating from the early years of cinema. According to the historian Ronald Hutton, the current state of the observance of Christmas is largely the result of a mid-Victorian revival of the holiday spearheaded by A Christmas Carol. Dickens catalysed the emerging Christmas as a family-centred festival of generosity, in contrast to the dwindling community-based and church-centred observations, as new middle-class expectations arose. Its archetypal figures (Scrooge, Tiny Tim, the Christmas ghosts) entered into Western cultural consciousness. "Merry Christmas", a prominent phrase from the tale, was popularised following the appearance of the story. The term Scrooge became a synonym for miser, and his exclamation "Bah! Humbug!'", a dismissal of the festive spirit, likewise gained currency as an idiom. The Victorian era novelist William Makepeace Thackeray called the book "a national benefit, and to every man and woman who reads it a personal kindness". Dickens was commemorated on the Series E £10 note issued by the Bank of England that circulated between 1992 and 2003. His portrait appeared on the reverse of the note accompanied by a scene from The Pickwick Papers. The Charles Dickens School is a high school in Broadstairs, Kent. A theme park, Dickens World, standing in part on the site of the former naval dockyard where Dickens's father once worked in the Navy Pay Office, opened in Chatham in 2007. To celebrate the 200th anniversary of the birth of Charles Dickens in 2012, the Museum of London held the UK's first major exhibition on the author in 40 years. In 2002, Dickens was number 41 in the BBC's poll of the 100 Greatest Britons. American literary critic Harold Bloom placed Dickens among the greatest Western writers of all time. In the 2003 UK survey The Big Read carried out by the BBC, five of Dickens's books were named in the Top 100. Actors who have portrayed Dickens on screen include Anthony Hopkins, Derek Jacobi, Simon Callow, Dan Stevens and Ralph Fiennes, the latter playing the author in The Invisible Woman (2013) which depicts Dickens's secret love affair with Ellen Ternan which lasted for thirteen years until his death in 1870. Dickens and his publications have appeared on a number of postage stamps in countries including: the United Kingdom (1970, 1993, 2011 and 2012 issued by the Royal Mail—their 2012 collection marked the bicentenary of Dickens's birth), the Soviet Union (1962), Antigua, Barbuda, Botswana, Cameroon, Dubai, Fujairah, St Lucia and Turks and Caicos Islands (1970), St Vincent (1987), Nevis (2007), Alderney, Gibraltar, Jersey and Pitcairn Islands (2012), Austria (2013), and Mozambique (2014). In 1976, a crater on the planet Mercury was named in his honour. In November 2018 it was reported that a previously lost portrait of a 31-year-old Dickens, by Margaret Gillies, had been found in Pietermaritzburg, South Africa. Gillies was an early supporter of women's suffrage and had painted the portrait in late 1843 when Dickens, aged 31, wrote A Christmas Carol. It was exhibited, to acclaim, at the Royal Academy of Arts in 1844. The Charles Dickens Museum is reported to have paid £180,000 for the portrait. Works Dickens published well over a dozen major novels and novellas, a large number of short stories, including a number of Christmas-themed stories, a handful of plays, and several non-fiction books. Novels Dickens's novels were initially serialised in weekly and monthly magazines, then reprinted in standard book formats. The Pickwick Papers (The Posthumous Papers of the Pickwick Club; monthly serial, April 1836 to November 1837) Oliver Twist (The Adventures of Oliver Twist; monthly serial in Bentley's Miscellany, February 1837 to April 1839) Nicholas Nickleby (The Life and Adventures of Nicholas Nickleby; monthly serial, April 1838 to October 1839) The Old Curiosity Shop (weekly serial in Master Humphrey's Clock, April 1840 to November 1841) Barnaby Rudge (Barnaby Rudge: A Tale of the Riots of Eighty; weekly serial in Master Humphrey's Clock, February to November 1841) A Christmas Carol (A Christmas Carol in Prose: Being a Ghost-story of Christmas; 1843) Martin Chuzzlewit (The Life and Adventures of Martin Chuzzlewit; monthly serial, January 1843 to July 1844) The Chimes (The Chimes: A Goblin Story of Some Bells That Rang an Old Year Out and a New Year In; 1844) The Cricket on the Hearth (The Cricket on the Hearth: A Fairy Tale of Home; 1845) The Battle of Life (The Battle of Life: A Love Story; 1846) Dombey and Son (Dealings with the Firm of Dombey and Son: Wholesale, Retail and for Exportation; monthly serial, October 1846 to April 1848) The Haunted Man (The Haunted Man and the Ghost's Bargain: A Fancy for Christmas-time; 1848) David Copperfield (The Personal History, Adventures, Experience and Observation of David Copperfield the Younger of Blunderstone Rookery [Which He Never Meant to Publish on Any Account]; monthly serial, May 1849 to November 1850) Bleak House (monthly serial, March 1852 to September 1853) Hard Times (Hard Times: For These Times; weekly serial in Household Words, 1 April 1854, to 12 August 1854) Little Dorrit (monthly serial, December 1855 to June 1857) A Tale of Two Cities (weekly serial in All the Year Round, 30 April 1859, to 26 November 1859) Great Expectations (weekly serial in All the Year Round, 1 December 1860 to 3 August 1861) Our Mutual Friend (monthly serial, May 1864 to November 1865) The Mystery of Edwin Drood (monthly serial, April 1870 to September 1870), left unfinished due to Dickens's death See also List of Dickensian characters Racism in the work of Charles Dickens Charles Dickens bibliography Notes References Sources Further reading Bradbury, Nicola, Charles Dickens' Great Expectations (St. Martin's Press, 1990) Douglas-Fairhurst, Robert, "Becoming Dickens 'The Invention of a Novelist'", London: Harvard University Press, 2011 Johnson, Edgar, Charles Dickens: his tragedy and triumph, New York: Simon and Schuster, 1952. In two volumes. Manning, Mick & Granström, Brita, Charles Dickens: Scenes From An Extraordinary Life, Frances Lincoln Children's Books, 2011. External links Works Charles Dickens's works on Bookwise Charles Dickens at the British Library Organisations and portals Charles Dickens on the Archives Hub Archival material at Leeds University Library The Dickens Fellowship, an international society dedicated to the study of Dickens and his Writings Correspondence of Charles Dickens, with related papers, ca. 1834–1955 Finding aid to Charles Dickens papers at Columbia University. Rare Book & Manuscript Library. Museums Dickens Museum Situated in a former Dickens House, 48 Doughty Street, London, WC1 Dickens Birthplace Museum Old Commercial Road, Portsmouth Victoria and Albert Museum The V&A's collections relating to Dickens Other Charles Dickens's Traveling Kit From the John Davis Batchelder Collection at the Library of Congress Charles Dickens's Walking Stick From the John Davis Batchelder Collection at the Library of Congress Charles Dickens Collection: First editions of Charles Dickens's works included in the Leonard Kebler gift (dispersed in the Division's collection). From the Rare Book and Special Collections Division at the Library of Congress 1812 births 1870 deaths 19th-century biographers 19th-century British newspaper founders 19th-century British non-fiction writers 19th-century British short story writers 19th-century English dramatists and playwrights 19th-century English non-fiction writers 19th-century English novelists 19th-century English poets 19th-century essayists 19th-century historians 19th-century British journalists 19th-century letter writers 19th-century British philanthropists 19th-century pseudonymous writers 19th-century travel writers Anglican writers British male essayists Burials at Westminster Abbey Children's rights activists Court reporters Critics of religions Critics of the Catholic Church Educational reformers English Anglicans English biographers English essayists English historians English historical novelists English letter writers English male dramatists and playwrights English male journalists English male non-fiction writers English male novelists English male poets English male short story writers English newspaper founders English philanthropists English prisoners and detainees English reformers English satirists English social commentators English travel writers Ghost story writers Lecturers Literacy and society theorists People from Camden Town People from Chatham, Kent People from Higham, Kent People from Somers Town, London Social reformers Trope theorists Victorian novelists Writers about activism and social change Writers from London Writers from Portsmouth Writers of Gothic fiction Writers of historical fiction set in the early modern period
6046
https://en.wikipedia.org/wiki/Cicero
Cicero
Marcus Tullius Cicero ( ; ; 3 January 106 BC – 7 December 43 BC) was a Roman statesman, lawyer, scholar, philosopher, writer and academic skeptic, who tried to uphold optimate principles during the political crises that led to the establishment of the Roman Empire. His extensive writings include treatises on rhetoric, philosophy and politics. He is considered one of Rome's greatest orators and prose stylists and the innovator of what became known as Ciceronian rhetoric. Cicero was educated in Rome and in Greece. He came from a wealthy municipal family of the Roman equestrian order, and served as consul in 63 BC. His influence on the Latin language was immense. He wrote more than three-quarters of extant Latin literature that is known to have existed in his lifetime, and it has been said that subsequent prose was either a reaction against or a return to his style, not only in Latin but in European languages up to the 19th century. Cicero introduced into Latin the arguments of the chief schools of Hellenistic philosophy and created a Latin philosophical vocabulary with neologisms such as , , , , and , distinguishing himself as a translator and philosopher. Though he was an accomplished orator and successful lawyer, Cicero believed his political career was his most important achievement. It was during his consulship that the Catiline conspiracy attempted to overthrow the government through an attack on the city by outside forces, and Cicero suppressed the revolt by summarily and controversially executing five conspirators without trial. During the chaotic middle period of the first century BC, marked by civil wars and the dictatorship of Julius Caesar, Cicero championed a return to the traditional republican government. Following Caesar's death, Cicero became an enemy of Mark Antony in the ensuing power struggle, attacking him in a series of speeches. He was proscribed as an enemy of the state by the Second Triumvirate and consequently executed by soldiers operating on their behalf in 43 BC having been intercepted during an attempted flight from the Italian peninsula. His severed hands and head were then, as a final revenge of Mark Antony, displayed on the Rostra. Petrarch's rediscovery of Cicero's letters is often credited for initiating the 14th-century Renaissance in public affairs, humanism, and classical Roman culture. According to Polish historian Tadeusz Zieliński, "the Renaissance was above all things a revival of Cicero, and only after him and through him of the rest of Classical antiquity." The peak of Cicero's authority and prestige came during the 18th-century Enlightenment, and his impact on leading Enlightenment thinkers and political theorists such as John Locke, David Hume, Montesquieu, and Edmund Burke was substantial. His works rank among the most influential in global culture, and today still constitute one of the most important bodies of primary material for the writing and revision of Roman history, especially the last days of the Roman Republic. Personal life Early life Marcus Tullius Cicero was born on 3 January 106 BC in Arpinum, a hill town southeast of Rome. He belonged to the tribus Cornelia. His father was a well-to-do member of the equestrian order and possessed good connections in Rome. However, being a semi-invalid, he could not enter public life and studied extensively to compensate. Although little is known about Cicero's mother, Helvia, it was common for the wives of important Roman citizens to be responsible for the management of the household. Cicero's brother Quintus wrote in a letter that she was a thrifty housewife. Cicero's cognomen, or personal surname, comes from the Latin for chickpea, . Plutarch explains that the name was originally given to one of Cicero's ancestors who had a cleft in the tip of his nose resembling a chickpea. However, it is more likely that Cicero's ancestors prospered through the cultivation and sale of chickpeas. Romans often chose down-to-earth personal surnames. The famous family names of Fabius, Lentulus, and Piso come from the Latin names of beans, lentils, and peas, respectively. Plutarch writes that Cicero was urged to change this deprecatory name when he entered politics, but refused, saying that he would make Cicero more glorious than Scaurus ("Swollen-ankled") and Catulus ("Puppy"). During this period in Roman history, "cultured" meant being able to speak both Latin and Greek. Cicero was therefore educated in the teachings of the ancient Greek philosophers, poets and historians; as he obtained much of his understanding of the theory and practice of rhetoric from the Greek poet Archias and from the Greek rhetorician Apollonius. Cicero used his knowledge of Greek to translate many of the theoretical concepts of Greek philosophy into Latin, thus translating Greek philosophical works for a larger audience. It was precisely his broad education that tied him to the traditional Roman elite. Cicero's interest in philosophy figured heavily in his later career and led to him providing a comprehensive account of Greek philosophy for a Roman audience, including creating a philosophical vocabulary in Latin. In 87 BC, Philo of Larissa, the head of the Platonic Academy that had been founded by Plato in Athens about 300 years earlier, arrived in Rome. Cicero, "inspired by an extraordinary zeal for philosophy", sat enthusiastically at his feet and absorbed Carneades' Academic Skeptic philosophy. Cicero said of Plato's Dialogues, that if Zeus were to speak, he would use their language. He would, in due course, honor them with his own convivial dialogues. According to Plutarch, Cicero was an extremely talented student, whose learning attracted attention from all over Rome, affording him the opportunity to study Roman law under Quintus Mucius Scaevola. Cicero's fellow students were Gaius Marius Minor, Servius Sulpicius Rufus (who became a famous lawyer, one of the few whom Cicero considered superior to himself in legal matters), and Titus Pomponius. The latter two became Cicero's friends for life, and Pomponius (who later received the nickname "Atticus", and whose sister married Cicero's brother) would become, in Cicero's own words, "as a second brother", with both maintaining a lifelong correspondence. In 79 BC, Cicero left for Greece, Asia Minor and Rhodes. This was perhaps to avoid the potential wrath of Sulla, as Plutarch claims, though Cicero himself says it was to hone his skills and improve his physical fitness. In Athens he studied philosophy with Antiochus of Ascalon, the 'Old Academic' and initiator of Middle Platonism. In Asia Minor, he met the leading orators of the region and continued to study with them. Cicero then journeyed to Rhodes to meet his former teacher, Apollonius Molon, who had previously taught him in Rome. Molon helped Cicero hone the excesses in his style, as well as train his body and lungs for the demands of public speaking. Charting a middle path between the competing Attic and Asiatic styles, Cicero would ultimately become considered second only to Demosthenes among history's orators. Family Cicero married Terentia probably at the age of 27, in 79 BC. According to the upper class mores of the day it was a marriage of convenience, but lasted harmoniously for nearly 30 years. Terentia's family was wealthy, probably the plebeian noble house of Terenti Varrones, thus meeting the needs of Cicero's political ambitions in both economic and social terms. She had a half-sister named Fabia, who as a child had become a Vestal Virgin, a great honour. Terentia was a strong willed woman and (citing Plutarch) "she took more interest in her husband's political career than she allowed him to take in household affairs." In the 50s BC, Cicero's letters to Terentia became shorter and colder. He complained to his friends that Terentia had betrayed him but did not specify in which sense. Perhaps the marriage could not outlast the strain of the political upheaval in Rome, Cicero's involvement in it, and various other disputes between the two. The divorce appears to have taken place in 51 BC or shortly before. In 46 or 45 BC, Cicero married a young girl, Publilia, who had been his ward. It is thought that Cicero needed her money, particularly after having to repay the dowry of Terentia, who came from a wealthy family. This marriage did not last long. Although his marriage to Terentia was one of convenience, it is commonly known that Cicero held great love for his daughter Tullia. When she suddenly became ill in February 45 BC and died after having seemingly recovered from giving birth to a son in January, Cicero was stunned. "I have lost the one thing that bound me to life" he wrote to Atticus. Atticus told him to come for a visit during the first weeks of his bereavement, so that he could comfort him when his pain was at its greatest. In Atticus's large library, Cicero read everything that the Greek philosophers had written about overcoming grief, "but my sorrow defeats all consolation." Caesar and Brutus as well as Servius Sulpicius Rufus sent him letters of condolence. Cicero hoped that his son Marcus would become a philosopher like him, but Marcus himself wished for a military career. He joined the army of Pompey in 49 BC and after Pompey's defeat at Pharsalus 48 BC, he was pardoned by Caesar. Cicero sent him to Athens to study as a disciple of the peripatetic philosopher Kratippos in 48 BC, but he used this absence from "his father's vigilant eye" to "eat, drink and be merry." After Cicero's death he joined the army of the Liberatores but was later pardoned by Augustus. Augustus's bad conscience for not having objected to Cicero's being put on the proscription list during the Second Triumvirate led him to aid considerably Marcus Minor's career. He became an augur, and was nominated consul in 30 BC together with Augustus. As such, he was responsible for revoking the honors of Mark Antony, who was responsible for the proscription, and could in this way take revenge. Later he was appointed proconsul of Syria and the province of Asia. Public career Early legal activity Cicero wanted to pursue a public career in politics along the steps of the . In 90–88 BC, he served both Pompeius Strabo and Lucius Cornelius Sulla as they campaigned in the Social War, though he had no taste for military life, being an intellectual first and foremost. Cicero started his career as a lawyer around 83–81 BC. The first extant speech is a private case from 81 BC (the ), delivered when Cicero was aged 26, though he refers throughout to previous defenses he had already undertaken. His first major public case, of which a written record is still extant, was his 80 BC defense of Sextus Roscius on the charge of patricide. Taking this case was a courageous move for Cicero; patricide was considered an appalling crime, and the people whom Cicero accused of the murder, the most notorious being Chrysogonus, were favorites of Sulla. At this time it would have been easy for Sulla to have the unknown Cicero murdered. Cicero's defense was an indirect challenge to the dictator Sulla, and on the strength of his case, Roscius was acquitted. Soon after, Cicero again challenged Sulla, by criticising his disenfranchisement of Italian towns in a lost speech on behalf of a woman from Arretium. Cicero's case in the was divided into three parts. The first part detailed exactly the charge brought by Ericius. Cicero explained how a rustic son of a farmer, who lives off the pleasures of his own land, would not have gained anything from committing patricide because he would have eventually inherited his father's land anyway. The second part concerned the boldness and greed of two of the accusers, Magnus and Capito. Cicero told the jury that they were the more likely perpetrators of murder because the two were greedy, both for conspiring together against a fellow kinsman and, in particular, Magnus, for his boldness and for being unashamed to appear in court to support the false charges. The third part explained that Chrysogonus had immense political power, and the accusation was successfully made due to that power. Even though Chrysogonus may not have been what Cicero said he was, through rhetoric Cicero successfully made him appear to be a foreign freed man who prospered by devious means in the aftermath of the civil war. Cicero surmised that it showed what kind of a person he was and that something like murder was not beneath him. Early political career His first office was as one of the twenty annual quaestors, a training post for serious public administration in a diversity of areas, but with a traditional emphasis on administration and rigorous accounting of public monies under the guidance of a senior magistrate or provincial commander. Cicero served as quaestor in western Sicily in 75 BC and demonstrated honesty and integrity in his dealings with the inhabitants. As a result, the grateful Sicilians asked Cicero to prosecute Gaius Verres, a governor of Sicily, who had plundered the province. His prosecution of Gaius Verres was a great forensic success for Cicero. Governor Gaius Verres hired the prominent lawyer of a noble family Quintus Hortensius Hortalus. After a lengthy period in Sicily collecting testimonials and evidence and persuading witnesses to come forward, Cicero returned to Rome and won the case in a series of dramatic court battles. His unique style of oratory set him apart from the flamboyant Hortensius. On the conclusion of this case, Cicero came to be considered the greatest orator in Rome. The view that Cicero may have taken the case for reasons of his own is viable. Hortensius was, at this point, known as the best lawyer in Rome; to beat him would guarantee much success and the prestige that Cicero needed to start his career. Cicero's oratorical ability is shown in his character assassination of Verres and various other techniques of persuasion used on the jury. One such example is found in the speech Against Verres I, where he states "with you on this bench, gentlemen, with Marcus Acilius Glabrio as your president, I do not understand what Verres can hope to achieve". Oratory was considered a great art in ancient Rome and an important tool for disseminating knowledge and promoting oneself in elections, in part because there were no regular newspapers or mass media. Cicero was neither a patrician nor a plebeian noble; his rise to political office despite his relatively humble origins has traditionally been attributed to his brilliance as an orator. Cicero grew up in a time of civil unrest and war. Sulla's victory in the first of a series of civil wars led to a new constitutional framework that undermined (liberty), the fundamental value of the Roman Republic. Nonetheless, Sulla's reforms strengthened the position of the equestrian class, contributing to that class's growing political power. Cicero was both an Italian and a , but more importantly he was a Roman constitutionalist. His social class and loyalty to the Republic ensured that he would "command the support and confidence of the people as well as the Italian middle classes". The optimates faction never truly accepted Cicero, and this undermined his efforts to reform the Republic while preserving the constitution. Nevertheless, he successfully ascended the cursus honorum, holding each magistracy at or near the youngest possible age: quaestor in 75 BC (age 30), aedile in 69 BC (age 36), and praetor in 66 BC (age 39), when he served as president of the "Reclamation" (or extortion) Court. He was then elected consul at age 42. Consulship Cicero, seizing the opportunity offered by optimate fear of reform, was elected consul for the year 63 BC; he was elected with the support of every unit of the centuriate assembly, rival members of the post-Sullan establishment, and the leaders of municipalities throughout post–Social War Italy. His co-consul for the year, Gaius Antonius Hybrida, played a minor role. He began his consular year by opposing a land bill proposed by a plebeian tribune which would have appointed commissioners with semi-permanent authority over land reform. Cicero was also active in the courts, defending Gaius Rabirius from accusations of participating in the unlawful killing of plebeian tribune Lucius Appuleius Saturninus in 100 BC. The prosecution occurred before the and threatened to reopen conflict between the Marian and Sullan factions at Rome. Cicero defended the use of force as being authorised by a , which would prove similar to his own use of force under such conditions. The Catilinarian Conspiracy Most famouslyin part because of his own publicityhe thwarted a conspiracy led by Lucius Sergius Catilina to overthrow the Roman Republic with the help of foreign armed forces. Cicero procured a senatus consultum ultimum (a recommendation from the senate attempting to legitimise the use of force) and drove Catiline from the city with four vehement speeches (the Catiline Orations), which remain outstanding examples of his rhetorical style. The Orations listed Catiline and his followers' debaucheries, and denounced Catiline's senatorial sympathizers as roguish and dissolute debtors clinging to Catiline as a final and desperate hope. Cicero demanded that Catiline and his followers leave the city. At the conclusion of Cicero's first speech (which was made in the Temple of Jupiter Stator), Catiline hurriedly left the Senate. In his following speeches, Cicero did not directly address Catiline. He delivered the second and third orations before the people, and the last one again before the Senate. By these speeches, Cicero wanted to prepare the Senate for the worst possible case; he also delivered more evidence, against Catiline. Catiline fled and left behind his followers to start the revolution from within while he himself assaulted the city with an army of "moral and financial bankrupts, or of honest fanatics and adventurers". It is alleged that Catiline had attempted to involve the Allobroges, a tribe of Transalpine Gaul, in their plot, but Cicero, working with the Gauls, was able to seize letters that incriminated the five conspirators and forced them to confess in front of the Senate. The senate then deliberated upon the conspirators' punishment. As it was the dominant advisory body to the various legislative assemblies rather than a judicial body, there were limits to its power; however, martial law was in effect, and it was feared that simple house arrest or exile – the standard options – would not remove the threat to the state. At first Decimus Junius Silanus spoke for the "extreme penalty"; many were swayed by Julius Caesar, who decried the precedent it would set and argued in favor of life imprisonment in various Italian towns. Cato the Younger rose in defense of the death penalty and the entire Senate finally agreed on the matter. Cicero had the conspirators taken to the Tullianum, the notorious Roman prison, where they were strangled. Cicero himself accompanied the former consul Publius Cornelius Lentulus Sura, one of the conspirators, to the Tullianum. Cicero received the honorific "pater patriae" for his efforts to suppress the conspiracy, but lived thereafter in fear of trial or exile for having put Roman citizens to death without trial. While the senatus consultum ultimum gave some legitimacy to the use of force against the conspirators, Cicero also argued that Catiline's conspiracy, by virtue of its treason, made the conspirators enemies of the state and forfeited the protections intrinsically possessed by Roman citizens. The consuls moved decisively. Antonius Hybrida was dispatched to defeat Catiline in battle that year, preventing Crassus or Pompey from exploiting the situation for their own political aims. After the suppression of the conspiracy, Cicero was proud of his accomplishment. Some of his political enemies argued that though the act gained Cicero popularity, he exaggerated the extent of his success. He overestimated his popularity again several years later after being exiled from Italy and then allowed back from exile. At this time, he claimed that the republic would be restored along with him. Many Romans at the time, led by Populares politicians Gaius Julius Caesar and patrician turned plebeian Publius Clodius Pulcher believed that Cicero's evidence against Catiline was fabricated and the witnesses were bribed. Cicero, who had been elected consul with the support of the Optimates, promoted their position as advocates of the status quo resisting social changes, especially more privileges for the average inhabitants of Rome. Shortly after completing his consulship, in late 62 BC, Cicero arranged the purchase of a large townhouse on the Palatine Hill previously owned by Rome's richest citizen, Marcus Licinius Crassus. It cost an exorbitant sum, 3.5 million sesterces, which required Cicero to arrange for a loan from his co-consul Gaius Antonius Hybrida based on the expected profits from Antonius's proconsulship in Macedonia. At the beginning of his consulship, Cicero had made an arrangement with Hybrida to grant Hybrida the profitable province of Macedonia that had been granted to Cicero by the Senate in exchange for Hybrida staying out of Cicero's way for the year and a quarter of the profits from the province. In return Cicero gained a lavish house which he proudly boasted was "in conspectu prope totius urbis" (in sight of nearly the whole city), only a short walk from the Roman Forum. Exile and return In 60 BC, Julius Caesar invited Cicero to be the fourth member of his existing partnership with Pompey and Marcus Licinius Crassus, an assembly that would eventually be called the First Triumvirate. Cicero refused the invitation because he suspected it would undermine the Republic. During Caesar's consulship of 59 BC, the triumvirate had achieved many of their goals of land reform, publicani debt forgiveness, ratification of Pompeian conquests, etc. With Caesar leaving for his provinces, they wished to maintain their hold on politics. They engineered the adoption of patrician Publius Clodius Pulcher into a plebeian family and had him elected as one of the ten tribunes of the plebs for 58 BC. Clodius used the triumvirate's backing to push through legislation that benefited them. He introduced several laws (the leges Clodiae) that made him popular with the people, strengthening his power base, then he turned on Cicero by threatening exile to anyone who executed a Roman citizen without a trial. Cicero, having executed members of the Catiline conspiracy four years previously without formal trial, was clearly the intended target. Furthermore, many believed that Clodius acted in concert with the triumvirate who feared that Cicero would seek to abolish many of Caesar's accomplishments while consul the year before. Cicero argued that the senatus consultum ultimum indemnified him from punishment, and he attempted to gain the support of the senators and consuls, especially of Pompey. Cicero grew out his hair, dressed in mourning and toured the streets. Clodius' gangs dogged him, hurling abuse, stones and even excrement. Hortensius, trying to rally to his old rival's support, was almost lynched. The Senate and the consuls were cowed. Caesar, who was still encamped near Rome, was apologetic but said he could do nothing when Cicero brought himself to grovel in the proconsul's tent. Everyone seemed to have abandoned Cicero. After Clodius passed a law to deny to Cicero fire and water (i.e. shelter) within four hundred miles of Rome, Cicero went into exile. He arrived at Thessalonica, on 23 May 58 BC. In his absence, Clodius, who lived next door to Cicero on the Palatine, arranged for Cicero's house to be confiscated by the state, and was even able to purchase a part of the property in order to extend his own house. After demolishing Cicero's house, Clodius had the land consecrated and symbolically erected a temple of Liberty (aedes Libertatis) on the vacant land. Cicero's exile caused him to fall into depression. He wrote to Atticus: "Your pleas have prevented me from committing suicide. But what is there to live for? Don't blame me for complaining. My afflictions surpass any you ever heard of earlier". After the intervention of recently elected tribune Titus Annius Milo, acting on the behalf of Pompey who wanted Cicero as a client, the Senate voted in favor of recalling Cicero from exile. Clodius cast the single vote against the decree. Cicero returned to Italy on 5 August 57 BC, landing at Brundisium. He was greeted by a cheering crowd, and, to his delight, his beloved daughter Tullia. In his Oratio De Domo Sua Ad Pontifices, Cicero convinced the College of Pontiffs to rule that the consecration of his land was invalid, thereby allowing him to regain his property and rebuild his house on the Palatine. Cicero tried to re-enter politics as an independent operator, but his attempts to attack portions of Caesar's legislation were unsuccessful and encouraged Caesar to re-solidify his political alliance with Pompey and Crassus. The conference at Luca in 56 BC left the three-man alliance in domination of the republic's politics; this forced Cicero to recant and support the triumvirate out of fear from being entirely excluded from public life. After the conference Cicero lavishly praised Caesar's achievements, got the Senate to vote a thanksgiving for Caesar's victories and grant money to pay his troops. He also delivered a speech 'On the consular provinces' () which checked an attempt by Caesar's enemies to strip him of his provinces in Gaul. After this, a cowed Cicero concentrated on his literary works. It is uncertain whether he was directly involved in politics for the following few years. Governorship of Cilicia In 51 BC he reluctantly accepted a promagistracy (as proconsul) in Cilicia for the year; there were few other former consuls eligible as a result of a legislative requirement enacted by Pompey in 52 BC specifying an interval of five years between a consulship or praetorship and a provincial command. He served as proconsul of Cilicia from May 51 BC, arriving in the provinces three months later around August. He was given instructions to keep nearby Cappadocia loyal to King Ariobarzanes III, which he achieved 'satisfactorily without war'. In 53 BC Marcus Licinius Crassus had been defeated by the Parthians at the Battle of Carrhae. This opened the Roman East for a Parthian invasion, causing unrest in Syria and Cilicia. Cicero restored calm by his mild system of government. He discovered that a great amount of public property had been embezzled by corrupt previous governors and members of their staff, and did his utmost to restore it. Thus he greatly improved the condition of the cities. He retained the civil rights of, and exempted from penalties, the men who gave the property back. Besides this, he was extremely frugal in his outlays for staff and private expenses during his governorship, and this made him highly popular among the natives. Previous governors had extorted enormous sums from the provincials in order to supply their households and bodyguards. Besides his activity in ameliorating the hard pecuniary situation of the province, Cicero was also creditably active in the military sphere. Early in his governorship he received information that prince Pacorus, son of Orodes II the king of the Parthians, had crossed the Euphrates, and was ravaging the Syrian countryside and had even besieged Cassius (the interim Roman commander in Syria) in Antioch. Cicero eventually marched with two understrength legions and a large contingent of auxiliary cavalry to Cassius's relief. Pacorus and his army had already given up on besieging Antioch and were heading south through Syria, ravaging the countryside again. Cassius and his legions followed them, harrying them wherever they went, eventually ambushing and defeating them near Antigonea. Another large troop of Parthian horsemen was defeated by Cicero's cavalry who happened to run into them while scouting ahead of the main army. Cicero next defeated some robbers who were based on Mount Amanus and was hailed as imperator by his troops. Afterwards he led his army against the independent Cilician mountain tribes, besieging their fortress of Pindenissum. It took him 47 days to reduce the place, which fell in December. On 30 July 50 BC Cicero left the province to his brother Quintus, who had accompanied him on his governorship as his legate. On his way back to Rome he stopped in Rhodes and then went to Athens, where he caught up with his old friend Titus Pomponius Atticus and met men of great learning. Julius Caesar's civil war Cicero arrived in Rome on 4 January 49 BC. He stayed outside the pomerium, to retain his promagisterial powers: either in expectation of a triumph or to retain his independent command authority in the coming civil war. The struggle between Pompey and Julius Caesar grew more intense in 50 BC. Cicero favored Pompey, seeing him as a defender of the senate and Republican tradition, but at that time avoided openly alienating Caesar. When Caesar invaded Italy in 49 BC, Cicero fled Rome. Caesar, seeking an endorsement by a senior senator, courted Cicero's favor, but even so Cicero slipped out of Italy and traveled to Dyrrachium (Epidamnos), Illyria, where Pompey's staff was situated. Cicero traveled with the Pompeian forces to Pharsalus in 48 BC, though he was quickly losing faith in the competence and righteousness of the Pompeian side. Eventually, he provoked the hostility of his fellow senator Cato, who told him that he would have been of more use to the cause of the optimates if he had stayed in Rome. After Caesar's victory at the Battle of Pharsalus on 9 August, Cicero refused to take command of the Pompeian forces and continue the war. He returned to Rome, still as a promagistrate with his lictors, in 47 BC, and dismissed them upon his crossing the pomerium and renouncing his command. Caesar pardoned him and Cicero tried to adjust to the situation and maintain his political work, hoping that Caesar might revive the Republic and its institutions. In a letter to Varro on , Cicero outlined his strategy under Caesar's dictatorship. Cicero, however, was taken by surprise when the Liberatores assassinated Caesar on the ides of March, 44 BC. Cicero was not included in the conspiracy, even though the conspirators were sure of his sympathy. Marcus Junius Brutus called out Cicero's name, asking him to restore the republic when he lifted his bloodstained dagger after the assassination. A letter Cicero wrote in February 43 BC to Trebonius, one of the conspirators, began, "How I could wish that you had invited me to that most glorious banquet on the Ides of March!" Cicero became a popular leader during the period of instability following the assassination. He had no respect for Mark Antony, who was scheming to take revenge upon Caesar's murderers. In exchange for amnesty for the assassins, he arranged for the Senate to agree not to declare Caesar to have been a tyrant, which allowed the Caesarians to have lawful support and kept Caesar's reforms and policies intact. Opposition to Mark Antony and death Cicero and Antony now became the two leading men in Rome: Cicero as spokesman for the Senate; Antony as consul, leader of the Caesarian faction, and unofficial executor of Caesar's public will. Relations between the two were never friendly and worsened after Cicero claimed that Antony was taking liberties in interpreting Caesar's wishes and intentions. Octavian was Caesar's adopted son and heir. After he returned to Italy, Cicero began to play him against Antony. He praised Octavian, declaring he would not make the same mistakes as his father. He attacked Antony in a series of speeches he called the Philippics, after Demosthenes's denunciations of Philip II of Macedon. At the time Cicero's popularity as a public figure was unrivalled. Cicero supported Decimus Junius Brutus Albinus as governor of Cisalpine Gaul (Gallia Cisalpina) and urged the Senate to name Antony an enemy of the state. The speech of Lucius Piso, Caesar's father-in-law, delayed proceedings against Antony. Antony was later declared an enemy of the state when he refused to lift the siege of Mutina, which was in the hands of Decimus Brutus. Cicero's plan to drive out Antony failed. Antony and Octavian reconciled and allied with Lepidus to form the Second Triumvirate after the successive battles of Forum Gallorum and Mutina. The Triumvirate began proscribing their enemies and potential rivals immediately after legislating the alliance into official existence for a term of five years with consular imperium. Cicero and all of his contacts and supporters were numbered among the enemies of the state, even though Octavian argued for two days against Cicero being added to the list. Cicero was one of the most viciously and doggedly hunted among the proscribed. He was viewed with sympathy by a large segment of the public and many people refused to report that they had seen him. He was caught on 7 December 43 BC leaving his villa in Formiae in a litter heading to the seaside, where he hoped to embark on a ship destined for Macedonia. When his killers – Herennius (a Centurion) and Popilius (a Tribune) – arrived, Cicero's own slaves said they had not seen him, but he was given away by Philologus, a freedman of his brother Quintus Cicero. As reported by Seneca the Elder, according to the historian Aufidius Bassus, Cicero's last words are said to have been: He bowed to his captors, leaning his head out of the litter in a gladiatorial gesture to ease the task. By baring his neck and throat to the soldiers, he was indicating that he would not resist. According to Plutarch, Herennius first slew him, then cut off his head. On Antony's instructions his hands, which had penned the Philippics against Antony, were cut off as well; these were nailed along with his head on the Rostra in the Forum Romanum according to the tradition of Marius and Sulla, both of whom had displayed the heads of their enemies in the Forum. Cicero was the only victim of the proscriptions who was displayed in that manner. According to Cassius Dio, in a story often mistakenly attributed to Plutarch, Antony's wife Fulvia took Cicero's head, pulled out his tongue, and jabbed it repeatedly with her hairpin in final revenge against Cicero's power of speech. Cicero's son, Marcus Tullius Cicero Minor, during his year as a consul in 30 BC, avenged his father's death, to a certain extent, when he announced to the Senate Mark Antony's naval defeat at Actium in 31 BC by Octavian. Octavian is reported to have praised Cicero as a patriot and a scholar of meaning in later times, within the circle of his family. However, it was Octavian's acquiescence that had allowed Cicero to be killed, as Cicero was condemned by the new triumvirate. Cicero's career as a statesman was marked by inconsistencies and a tendency to shift his position in response to changes in the political climate. His indecision may be attributed to his sensitive and impressionable personality; he was prone to overreaction in the face of political and private change. "Would that he had been able to endure prosperity with greater self-control, and adversity with more fortitude!" wrote C. Asinius Pollio, a contemporary Roman statesman and historian. Legacy Cicero has been traditionally considered the master of Latin prose, with Quintilian declaring that Cicero was "not the name of a man, but of eloquence itself." The English words Ciceronian (meaning "eloquent") and cicerone (meaning "local guide") derive from his name. He is credited with transforming Latin from a modest utilitarian language into a versatile literary medium capable of expressing abstract and complicated thoughts with clarity. Julius Caesar praised Cicero's achievement by saying "it is more important to have greatly extended the frontiers of the Roman spirit than the frontiers of the Roman empire". According to John William Mackail, "Cicero's unique and imperishable glory is that he created the language of the civilized world, and used that language to create a style which nineteen centuries have not replaced, and in some respects have hardly altered." Cicero was also an energetic writer with an interest in a wide variety of subjects, in keeping with the Hellenistic philosophical and rhetorical traditions in which he was trained. The quality and ready accessibility of Ciceronian texts favored very wide distribution and inclusion in teaching curricula, as suggested by a graffito at Pompeii, admonishing: "You will like Cicero, or you will be whipped". Cicero was greatly admired by influential Church Fathers such as Augustine of Hippo, who credited Cicero's lost Hortensius for his eventual conversion to Christianity, and St. Jerome, who had a feverish vision in which he was accused of being "follower of Cicero and not of Christ" before the judgment seat. This influence further increased after the Early Middle Ages in Europe, which more of his writings survived than any other Latin author. Medieval philosophers were influenced by Cicero's writings on natural law and innate rights. Petrarch's rediscovery of Cicero's letters provided the impetus for searches for ancient Greek and Latin writings scattered throughout European monasteries, and the subsequent rediscovery of classical antiquity led to the Renaissance. Subsequently, Cicero became synonymous with classical Latin to such an extent that a number of humanist scholars began to assert that no Latin word or phrase should be used unless it appeared in Cicero's works, a stance criticised by Erasmus. His voluminous correspondence, much of it addressed to his friend Atticus, has been especially influential, introducing the art of refined letter writing to European culture. Cornelius Nepos, the first century BC biographer of Atticus, remarked that Cicero's letters contained such a wealth of detail "concerning the inclinations of leading men, the faults of the generals, and the revolutions in the government" that their reader had little need for a history of the period. Among Cicero's admirers were Desiderius Erasmus, Martin Luther, and John Locke. Following the invention of Johannes Gutenberg's printing press, De Officiis was the second book printed in Europe, after the Gutenberg Bible. Scholars note Cicero's influence on the rebirth of religious toleration in the 17th century. Cicero was especially popular with the Philosophes of the 18th century, including Edward Gibbon, Diderot, David Hume, Montesquieu, and Voltaire. Gibbon wrote of his first experience reading the author's collective works thus: "I tasted the beauty of the language; I breathed the spirit of freedom; and I imbibed from his precepts and examples the public and private sense of a man...after finishing the great author, a library of eloquence and reason, I formed a more extensive plan of reviewing the Latin classics..." Voltaire called Cicero "the greatest as well as the most elegant of Roman philosophers" and even staged a play based on Cicero's role in the Catilinarian conspiracy, called Rome Sauvée, ou Catilina, to "make young people who go to the theatre acquainted with Cicero." Voltaire was spurred to pen the drama as a rebuff to his rival Claude Prosper Jolyot de Crébillon's own play Catilina, which had portrayed Cicero as a coward and villain who hypocritically married his own daughter to Catiline. Montesquieu produced his "Discourse on Cicero" in 1717, in which he heaped praise on the author because he rescued "philosophy from the hands of scholars, and freed it from the confusion of a foreign language". Montesquieu went on to declare that Cicero was "of all the ancients, the one who had the most personal merit, and whom I would prefer to resemble." Internationally, Cicero the republican inspired the Founding Fathers of the United States and the revolutionaries of the French Revolution. John Adams said, "As all the ages of the world have not produced a greater statesman and philosopher united than Cicero, his authority should have great weight." Jefferson names Cicero as one of a handful of major figures who contributed to a tradition "of public right" that informed his draft of the Declaration of Independence and shaped American understandings of "the common sense" basis for the right of revolution. Camille Desmoulins said of the French republicans in 1789 that they were "mostly young people who, nourished by the reading of Cicero at school, had become passionate enthusiasts for liberty". Jim Powell starts his book on the history of liberty with the sentence: "Marcus Tullius Cicero expressed principles that became the bedrock of liberty in the modern world." Likewise, no other ancient personality has inspired as much venomous dislike as Cicero, especially in more modern times. His commitment to the values of the Republic accommodated a hatred of the poor and persistent opposition to the advocates and mechanisms of popular representation. Friedrich Engels referred to him as "the most contemptible scoundrel in history" for upholding republican "democracy" while at the same time denouncing land and class reforms. Cicero has faced criticism for exaggerating the democratic qualities of republican Rome, and for defending the Roman oligarchy against the popular reforms of Caesar. Michael Parenti admits Cicero's abilities as an orator, but finds him a vain, pompous and hypocritical personality who, when it suited him, could show public support for popular causes that he privately despised. Parenti presents Cicero's prosecution of the Catiline conspiracy as legally flawed at least, and possibly unlawful. Cicero also had an influence on modern astronomy. Nicolaus Copernicus, searching for ancient views on earth motion, said that he "first ... found in Cicero that Hicetas supposed the earth to move." Notably, "Cicero" was the name attributed to size 12 font in typesetting table drawers. For ease of reference, type sizes 3, 4, 5, 6, 7, 8, 9, 10, 12, 14, 16, and 20 were all given different names. Works Cicero was declared a righteous pagan by the Early Church, and therefore many of his works were deemed worthy of preservation. Subsequent Roman and medieval Christian writers quoted liberally from his works De re publica (On the Commonwealth) and De Legibus (On the Laws), and much of his work has been recreated from these surviving fragments. Cicero also articulated an early, abstract conceptualization of rights, based on ancient law and custom. Of Cicero's books, six on rhetoric have survived, as well as parts of seven on philosophy. Of his speeches, 88 were recorded, but only 52 survive. In archaeology Cicero's great repute in Italy has led to numerous ruins being identified as having belonged to him, though none have been substantiated with absolute certainty. In Formia, two Roman-era ruins are popularly believed to be Cicero's mausoleum, the Tomba di Cicerone, and the villa where he was assassinated in 43 BC. The latter building is centered around a central hall with Doric columns and a coffered vault, with a separate nymphaeum, on five acres of land near Formia. A modern villa was built on the site after the Rubino family purchased the land from Ferdinand II of the Two Sicilies in 1868. Cicero's supposed tomb is a 24-meter (79 feet) tall tower on an opus quadratum base on the ancient Via Appia outside of Formia. Some suggest that it is not in fact Cicero's tomb, but a monument built on the spot where Cicero was intercepted and assassinated while trying to reach the sea. In Pompeii, a large villa excavated in the mid 18th century just outside the Herculaneum Gate was widely believed to have been Cicero's, who was known to have owned a holiday villa in Pompeii he called his Pompeianum. The villa was stripped of its fine frescoes and mosaics and then re-buried after 1763 – it has yet to be re-excavated. However, contemporaneous descriptions of the building from the excavators combined with Cicero's own references to his Pompeianum differ, making it unlikely that it is Cicero's villa. In Rome, the location of Cicero's house has been roughly identified from excavations of the Republican-era stratum on the northwestern slope of the Palatine Hill. Cicero's domus has long been known to have stood in the area, according to his own descriptions and those of later authors, but there is some debate about whether it stood near the base of the hill, very close to the Roman Forum, or nearer to the summit. During his life the area was the most desirable in Rome, densely occupied with Patrician houses including the Domus Publica of Julius Caesar and the home of Cicero's mortal enemy Clodius. Notable fictional portrayals In Dante's 1320 poem the Divine Comedy, the author encounters Cicero, among other philosophers, in Limbo. Ben Jonson dramatised the conspiracy of Catiline in his play Catiline His Conspiracy, featuring Cicero as a character. Cicero also appears as a minor character in William Shakespeare's play Julius Caesar. Cicero was portrayed on the motion picture screen by British actor Alan Napier in the 1953 film Julius Caesar, based on Shakespeare's play. He has also been played by such noted actors as Michael Hordern (in Cleopatra), and André Morell (in the 1970 Julius Caesar). Most recently, Cicero was portrayed by David Bamber in the HBO series Rome (2005–2007) and appeared in both seasons. In the historical novel series Masters of Rome, Colleen McCullough presents a not-so-flattering depiction of Cicero's career, showing him struggling with an inferiority complex and vanity, morally flexible and fatally indiscreet, while his rival Julius Caesar is shown in a more approving light. Cicero is portrayed as a hero in the novel A Pillar of Iron by Taylor Caldwell (1965). Robert Harris' novels Imperium, Lustrum (published under the name Conspirata in the United States) and Dictator comprise a three-part series based on the life of Cicero. In these novels Cicero's character is depicted in a more favorable way than in those of McCullough, with his positive traits equaling or outweighing his weaknesses (while conversely Caesar is depicted as more sinister than in McCullough). Cicero is a major recurring character in the Roma Sub Rosa series of mystery novels by Steven Saylor. He also appears several times as a peripheral character in John Maddox Roberts' SPQR series. Samuel Barnett portrays Cicero in a 2017 audio drama series pilot produced by Big Finish Productions. A full series was released the following year. All episodes are written by David Llewellyn and directed and produced by Scott Handcock. See also Caecilia Attica Caecilia Metella (daughter of Metellus Celer) Civis romanus sum Clausula (rhetoric) A Dialogue Concerning Oratorical Partitions E pluribus unum Esse quam videri Ipse dixit List of ancient Romans Lorem ipsum Marcantonius Majoragio Marcus Tullius Tiro Marius Nizolius O tempora, o mores! Otium Socratici viri Tempest in a teapot Translation Writings of Cicero Notes References Citations Sources Badian, E: "Cicero and the Commission of 146 B.C.", Collection Latomus 101 (1969), 54–65. Cicero, Marcus Tullius, Cicero's letters to Atticus, Vol, I, II, IV, VI, Cambridge University Press, Great Britain, 1965 Cicero, Marcus Tullius, Latin extracts of Cicero on Himself, translated by Charles Gordon Cooper, University of Queensland Press, Brisbane, 1963 Cicero, Marcus Tullius, Selected Political Speeches, Penguin Books Ltd, Great Britain, 1969 Cicero, Marcus Tullius, De Officiis (On Duties), translated by Walter Miller. Harvard University Press, 1913, Cicero, Marcus Tullius, Selected Works, Penguin Books Ltd, Great Britain, 1971 Cowell, F.R. (1948). Cicero and the Roman Republic. Penguin Books Plutarch Penguins Classics English translation by Rex Warner, Fall of the Roman Republic, Six Lives by Plutarch: Marius, Sulla, Crassus, Pompey, Caesar, Cicero (Penguin Books, 1958; with Introduction and notes by Robin Seager, 1972) Rawson, Beryl: The Politics of Friendship: Pompey and Cicero (Sydney University Press, 1978) Scullard, H.H. From the Gracchi to Nero, University Paperbacks, Great Britain, 1968 Smith, R.E: Cicero the Statesman (Cambridge University Press, 1966) Stockton, David: Cicero: A Political Biography (Oxford University Press, 1971) Uttschenko, Sergej L. (1978): Cicero, translated from Russian by Rosemarie Pattloch, VEB Deutscher Verlag der Wissenschaften, Berlin, Germany.Cicero Further reading Boissier, Gaston, Cicéron et ses amis. Étude sur la société romaine du temps de César (1884) Gildenhard, Ingo (2011). Creative Eloquence: The Construction of Reality in Cicero's Speeches. Oxford/New York: Oxford University Press. Hamza, Gabor, L'optimus status civitatis di Cicerone e la sua tradizione nel pensiero politico. In: Tradizione romanistica e Costituzione. Cinquanta anni della Corte Costituzionale della Repubblica Italiana. vol. II. Napoli, 2006. 1455–1468. Hamza, Gabor, Ciceros Verhältnis zu seinen Quellen, mit besonderer Berücksichtigung der Darstellung der Staatslehre in De re publica. KLIO – Beiträge zur alten Geschichte 67 (1985) 492–497. Hamza, Gabor, Cicero und der Idealtypus des iurisconsultus. Helixon 22–27 (1982–1987) 281–296. Hamza, Gabor, Il potere (lo Stato) nel pensiero di Cicerone e la sua attualità. Revista Internacional de Derecho Romano (RIDROM) 10 (2013) 1–25. Revista Internacional de Derecho Romano – Index Hamza, Gabor, Zur Interpretation des Naturrechts in den Werken von Cicero. Pázmány Law Review 2 (2014) 5–15. Treggiari, S. (2007). Terentia, Tullia and Publilia. The women of Cicero's family. London: Routledge. External links Works by Cicero Works by Cicero at Perseus Digital Library Works by Cicero at the Stoic Therapy eLibrary The Latin Library (Latin): Works of Cicero Dickinson College Commentaries: Against Verres 2.1.53–86 Dickinson College Commentaries: On Pompey's Command (De Imperio) 27–49 Horace MS 1b Laelius de Amicitia at OPenn Lewis E 66 Epistolae ad familiares (Letters to friends) Biographies and descriptions of Cicero's time Plutarch's biography of Cicero contained in the Parallel Lives Life of Cicero by Anthony Trollope, Volume I & Volume II Cicero by Rev. W. Lucas Collins (Ancient Classics for English Readers) Roman life in the days of Cicero by Rev. Alfred J. Church Social life at Rome in the Age of Cicero by W. Warde Fowler Dryden's translation of Cicero from Plutarch's Parallel Lives At Middlebury College website 106 BC births 43 BC deaths 2nd-century BC Romans 1st-century BC executions 1st-century BC philosophers 1st-century BC Roman augurs 1st-century BC Roman consuls 1st-century BC Latin writers Ancient landowners Ancient Roman equites Ancient Roman exiles Ancient Roman jurists Ancient Roman rhetoricians Ancient Roman scholars of religion Classical humanists Deaths by blade weapons Executed ancient Roman people Executed philosophers Executed writers Golden Age Latin writers Latin letter writers Optimates People executed by the Roman Republic People from Arpino Plebeian aediles Philosophers of Roman Italy Recipients of ancient Roman pardons Roman governors of Cilicia Roman quaestors Roman Republican praetors Roman-era Skeptic philosophers Roman-era students in Athens Translation scholars Translators of Ancient Greek texts Trope theorists Tullii
6233
https://en.wikipedia.org/wiki/Connected%20space
Connected space
In topology and related branches of mathematics, a connected space is a topological space that cannot be represented as the union of two or more disjoint non-empty open subsets. Connectedness is one of the principal topological properties that are used to distinguish topological spaces. A subset of a topological space is a if it is a connected space when viewed as a subspace of . Some related but stronger conditions are path connected, simply connected, and -connected. Another related notion is locally connected, which neither implies nor follows from connectedness. Formal definition A topological space is said to be if it is the union of two disjoint non-empty open sets. Otherwise, is said to be connected. A subset of a topological space is said to be connected if it is connected under its subspace topology. Some authors exclude the empty set (with its unique topology) as a connected space, but this article does not follow that practice. For a topological space the following conditions are equivalent: is connected, that is, it cannot be divided into two disjoint non-empty open sets. The only subsets of which are both open and closed (clopen sets) are and the empty set. The only subsets of with empty boundary are and the empty set. cannot be written as the union of two non-empty separated sets (sets for which each is disjoint from the other's closure). All continuous functions from to are constant, where is the two-point space endowed with the discrete topology. Historically this modern formulation of the notion of connectedness (in terms of no partition of into two separated sets) first appeared (independently) with N.J. Lennes, Frigyes Riesz, and Felix Hausdorff at the beginning of the 20th century. See for details. Connected components Given some point in a topological space the union of any collection of connected subsets such that each contains will once again be a connected subset. The connected component of a point in is the union of all connected subsets of that contain it is the unique largest (with respect to ) connected subset of that contains The maximal connected subsets (ordered by inclusion ) of a non-empty topological space are called the connected components of the space. The components of any topological space form a partition of : they are disjoint, non-empty and their union is the whole space. Every component is a closed subset of the original space. It follows that, in the case where their number is finite, each component is also an open subset. However, if their number is infinite, this might not be the case; for instance, the connected components of the set of the rational numbers are the one-point sets (singletons), which are not open. Proof: Any two distinct rational numbers are in different components. Take an irrational number and then set and Then is a separation of and . Thus each component is a one-point set. Let be the connected component of in a topological space and be the intersection of all clopen sets containing (called quasi-component of ) Then where the equality holds if is compact Hausdorff or locally connected. Disconnected spaces A space in which all components are one-point sets is called . Related to this property, a space is called if, for any two distinct elements and of , there exist disjoint open sets containing and containing such that is the union of and . Clearly, any totally separated space is totally disconnected, but the converse does not hold. For example take two copies of the rational numbers , and identify them at every point except zero. The resulting space, with the quotient topology, is totally disconnected. However, by considering the two copies of zero, one sees that the space is not totally separated. In fact, it is not even Hausdorff, and the condition of being totally separated is strictly stronger than the condition of being Hausdorff. Examples The closed interval in the standard subspace topology is connected; although it can, for example, be written as the union of and the second set is not open in the chosen topology of The union of and is disconnected; both of these intervals are open in the standard topological space is disconnected. A convex subset of is connected; it is actually simply connected. A Euclidean plane excluding the origin, is connected, but is not simply connected. The three-dimensional Euclidean space without the origin is connected, and even simply connected. In contrast, the one-dimensional Euclidean space without the origin is not connected. A Euclidean plane with a straight line removed is not connected since it consists of two half-planes. , the space of real numbers with the usual topology, is connected. The Sorgenfrey line is disconnected. If even a single point is removed from , the remainder is disconnected. However, if even a countable infinity of points are removed from , where the remainder is connected. If , then remains simply connected after removal of countably many points. Any topological vector space, e.g. any Hilbert space or Banach space, over a connected field (such as or ), is simply connected. Every discrete topological space with at least two elements is disconnected, in fact such a space is totally disconnected. The simplest example is the discrete two-point space. On the other hand, a finite set might be connected. For example, the spectrum of a discrete valuation ring consists of two points and is connected. It is an example of a Sierpiński space. The Cantor set is totally disconnected; since the set contains uncountably many points, it has uncountably many components. If a space is homotopy equivalent to a connected space, then is itself connected. The topologist's sine curve is an example of a set that is connected but is neither path connected nor locally connected. The general linear group (that is, the group of -by- real, invertible matrices) consists of two connected components: the one with matrices of positive determinant and the other of negative determinant. In particular, it is not connected. In contrast, is connected. More generally, the set of invertible bounded operators on a complex Hilbert space is connected. The spectra of commutative local ring and integral domains are connected. More generally, the following are equivalent The spectrum of a commutative ring is connected Every finitely generated projective module over has constant rank. has no idempotent (i.e., is not a product of two rings in a nontrivial way). An example of a space that is not connected is a plane with an infinite line deleted from it. Other examples of disconnected spaces (that is, spaces which are not connected) include the plane with an annulus removed, as well as the union of two disjoint closed disks, where all examples of this paragraph bear the subspace topology induced by two-dimensional Euclidean space. Path connectedness A is a stronger notion of connectedness, requiring the structure of a path. A path from a point to a point in a topological space is a continuous function from the unit interval to with and . A of is an equivalence class of under the equivalence relation which makes equivalent to if there is a path from to . The space is said to be path-connected (or pathwise connected or -connected) if there is exactly one path-component, i.e. if there is a path joining any two points in . Again, many authors exclude the empty space (by this definition, however, the empty space is not path-connected because it has zero path-components; there is a unique equivalence relation on the empty set which has zero equivalence classes). Every path-connected space is connected. The converse is not always true: examples of connected spaces that are not path-connected include the extended long line and the topologist's sine curve. Subsets of the real line are connected if and only if they are path-connected; these subsets are the intervals of . Also, open subsets of or are connected if and only if they are path-connected. Additionally, connectedness and path-connectedness are the same for finite topological spaces. Arc connectedness A space is said to be arc-connected or arcwise connected if any two topologically distinguishable points can be joined by an arc, which is an embedding . An arc-component of is a maximal arc-connected subset of ; or equivalently an equivalence class of the equivalence relation of whether two points can be joined by an arc or by a path whose points are topologically indistinguishable. Every Hausdorff space that is path-connected is also arc-connected; more generally this is true for a -Hausdorff space, which is a space where each image of a path is closed. An example of a space which is path-connected but not arc-connected is given by the line with two origins; its two copies of can be connected by a path but not by an arc. Intuition for path-connected spaces does not readily transfer to arc-connected spaces. Let be the line with two origins. The following are facts whose analogues hold for path-connected spaces, but do not hold for arc-connected spaces: Continuous image of arc-connected space may not be arc-connected: for example, a quotient map from an arc-connected space to its quotient with countably many (at least 2) topologically distinguishable points cannot be arc-connected due to too small cardinality. Arc-components may not be disjoint. For example, has two overlapping arc-components. Arc-connected product space may not be a product of arc-connected spaces. For example, is arc-connected, but is not. Arc-components of a product space may not be products of arc-components of the marginal spaces. For example, has a single arc-component, but has two arc-components. If arc-connected subsets have a non-empty intersection, then their union may not be arc-connected. For example, the arc-components of intersect, but their union is not arc-connected. Local connectedness A topological space is said to be locally connected at a point if every neighbourhood of contains a connected open neighbourhood. It is locally connected if it has a base of connected sets. It can be shown that a space is locally connected if and only if every component of every open set of is open. Similarly, a topological space is said to be if it has a base of path-connected sets. An open subset of a locally path-connected space is connected if and only if it is path-connected. This generalizes the earlier statement about and , each of which is locally path-connected. More generally, any topological manifold is locally path-connected. Locally connected does not imply connected, nor does locally path-connected imply path connected. A simple example of a locally connected (and locally path-connected) space that is not connected (or path-connected) is the union of two separated intervals in , such as . A classical example of a connected space that is not locally connected is the so called topologist's sine curve, defined as , with the Euclidean topology induced by inclusion in . Set operations The intersection of connected sets is not necessarily connected. The union of connected sets is not necessarily connected, as can be seen by considering . Each ellipse is a connected set, but the union is not connected, since it can be partitioned to two disjoint open sets and . This means that, if the union is disconnected, then the collection can be partitioned to two sub-collections, such that the unions of the sub-collections are disjoint and open in (see picture). This implies that in several cases, a union of connected sets necessarily connected. In particular: If the common intersection of all sets is not empty (), then obviously they cannot be partitioned to collections with disjoint unions. Hence the union of connected sets with non-empty intersection is connected. If the intersection of each pair of sets is not empty () then again they cannot be partitioned to collections with disjoint unions, so their union must be connected. If the sets can be ordered as a "linked chain", i.e. indexed by integer indices and , then again their union must be connected. If the sets are pairwise-disjoint and the quotient space is connected, then must be connected. Otherwise, if is a separation of then is a separation of the quotient space (since are disjoint and open in the quotient space). The set difference of connected sets is not necessarily connected. However, if and their difference is disconnected (and thus can be written as a union of two open sets and ), then the union of with each such component is connected (i.e. is connected for all ). Theorems Main theorem of connectedness: Let and be topological spaces and let be a continuous function. If is (path-)connected then the image is (path-)connected. This result can be considered a generalization of the intermediate value theorem. Every path-connected space is connected. Every locally path-connected space is locally connected. A locally path-connected space is path-connected if and only if it is connected. The closure of a connected subset is connected. Furthermore, any subset between a connected subset and its closure is connected. The connected components are always closed (but in general not open) The connected components of a locally connected space are also open. The connected components of a space are disjoint unions of the path-connected components (which in general are neither open nor closed). Every quotient of a connected (resp. locally connected, path-connected, locally path-connected) space is connected (resp. locally connected, path-connected, locally path-connected). Every product of a family of connected (resp. path-connected) spaces is connected (resp. path-connected). Every open subset of a locally connected (resp. locally path-connected) space is locally connected (resp. locally path-connected). Every manifold is locally path-connected. Arc-wise connected space is path connected, but path-wise connected space may not be arc-wise connected Continuous image of arc-wise connected set is arc-wise connected. Graphs Graphs have path connected subsets, namely those subsets for which every pair of points has a path of edges joining them. But it is not always possible to find a topology on the set of points which induces the same connected sets. The 5-cycle graph (and any -cycle with odd) is one such example. As a consequence, a notion of connectedness can be formulated independently of the topology on a space. To wit, there is a category of connective spaces consisting of sets with collections of connected subsets satisfying connectivity axioms; their morphisms are those functions which map connected sets to connected sets . Topological spaces and graphs are special cases of connective spaces; indeed, the finite connective spaces are precisely the finite graphs. However, every graph can be canonically made into a topological space, by treating vertices as points and edges as copies of the unit interval (see topological graph theory#Graphs as topological spaces). Then one can show that the graph is connected (in the graph theoretical sense) if and only if it is connected as a topological space. Stronger forms of connectedness There are stronger forms of connectedness for topological spaces, for instance: If there exist no two disjoint non-empty open sets in a topological space , must be connected, and thus hyperconnected spaces are also connected. Since a simply connected space is, by definition, also required to be path connected, any simply connected space is also connected. If the "path connectedness" requirement is dropped from the definition of simple connectivity, a simply connected space does not need to be connected. Yet stronger versions of connectivity include the notion of a contractible space. Every contractible space is path connected and thus also connected. In general, any path connected space must be connected but there exist connected spaces that are not path connected. The deleted comb space furnishes such an example, as does the above-mentioned topologist's sine curve. See also References Further reading . General topology Properties of topological spaces
6424
https://en.wikipedia.org/wiki/Corona%20Australis
Corona Australis
Corona Australis is a constellation in the Southern Celestial Hemisphere. Its Latin name means "southern crown", and it is the southern counterpart of Corona Borealis, the northern crown. It is one of the 48 constellations listed by the 2nd-century astronomer Ptolemy, and it remains one of the 88 modern constellations. The Ancient Greeks saw Corona Australis as a wreath rather than a crown and associated it with Sagittarius or Centaurus. Other cultures have likened the pattern to a turtle, ostrich nest, a tent, or even a hut belonging to a rock hyrax. Although fainter than its northern counterpart, the oval- or horseshoe-shaped pattern of its brighter stars renders it distinctive. Alpha and Beta Coronae Australis are the two brightest stars with an apparent magnitude of around 4.1. Epsilon Coronae Australis is the brightest example of a W Ursae Majoris variable in the southern sky. Lying alongside the Milky Way, Corona Australis contains one of the closest star-forming regions to the Solar System—a dusty dark nebula known as the Corona Australis Molecular Cloud, lying about 430 light years away. Within it are stars at the earliest stages of their lifespan. The variable stars R and TY Coronae Australis light up parts of the nebula, which varies in brightness accordingly. Name The name of the constellation was entered as "Corona Australis" when the International Astronomical Union (IAU) established the 88 modern constellations in 1922. In 1932, the name was instead recorded as "Corona Austrina" when the IAU's commission on notation approved a list of four-letter abbreviations for the constellations. The four-letter abbreviations were repealed in 1955. The IAU presently uses "Corona Australis" exclusively. Characteristics Corona Australis is a small constellation bordered by Sagittarius to the north, Scorpius to the west, Telescopium to the south, and Ara to the southwest. The three-letter abbreviation for the constellation, as adopted by the International Astronomical Union in 1922, is "CrA". The official constellation boundaries, as set by Belgian astronomer Eugène Delporte in 1930, are defined by a polygon of four segments (illustrated in infobox). In the equatorial coordinate system, the right ascension coordinates of these borders lie between and , while the declination coordinates are between −36.77° and −45.52°. Covering 128 square degrees, Corona Australis culminates at midnight around the 30th of June and ranks 80th in area. Only visible at latitudes south of 53° north, Corona Australis cannot be seen from the British Isles as it lies too far south, but it can be seen from southern Europe and readily from the southern United States. Features While not a bright constellation, Corona Australis is nonetheless distinctive due to its easily identifiable pattern of stars, which has been described as horseshoe- or oval-shaped. Though it has no stars brighter than 4th magnitude, it still has 21 stars visible to the unaided eye (brighter than magnitude 5.5). Nicolas Louis de Lacaille used the Greek letters Alpha through to Lambda to label the most prominent eleven stars in the constellation, designating two stars as Eta and omitting Iota altogether. Mu Coronae Australis, a yellow star of spectral type G5.5III and apparent magnitude 5.21, was labelled by Johann Elert Bode and retained by Benjamin Gould, who deemed it bright enough to warrant naming. Stars The only star in the constellation to have received a name is Alfecca Meridiana or Alpha CrA. The name combines the Arabic name of the constellation with the Latin for "southern". In Arabic, Alfecca means "break", and refers to the shape of both Corona Australis and Corona Borealis. Also called simply "Meridiana", it is a white main sequence star located 125 light years away from Earth, with an apparent magnitude of 4.10 and spectral type A2Va. A rapidly rotating star, it spins at almost 200 km per second at its equator, making a complete revolution in around 14 hours. Like the star Vega, it has excess infrared radiation, which indicates it may be ringed by a disk of dust. It is currently a main-sequence star, but will eventually evolve into a white dwarf; currently, it has a luminosity 31 times greater, and a radius and mass of 2.3 times that of the Sun. Beta Coronae Australis is an orange giant 474 light years from Earth. Its spectral type is K0II, and it is of apparent magnitude 4.11. Since its formation, it has evolved from a B-type star to a K-type star. Its luminosity class places it as a bright giant; its luminosity is 730 times that of the Sun, designating it one of the highest-luminosity K0-type stars visible to the naked eye. 100 million years old, it has a radius of 43 solar radii () and a mass of between 4.5 and 5 solar masses (). Alpha and Beta are so similar as to be indistinguishable in brightness to the naked eye. Some of the more prominent double stars include Gamma Coronae Australis—a pair of yellowish white stars 58 light years away from Earth, which orbit each other every 122 years. Widening since 1990, the two stars can be seen as separate with a 100 mm aperture telescope; they are separated by 1.3 arcseconds at an angle of 61 degrees. They have a combined visual magnitude of 4.2; each component is an F8V dwarf star with a magnitude of 5.01. Epsilon Coronae Australis is an eclipsing binary belonging to a class of stars known as W Ursae Majoris variables. These star systems are known as contact binaries as the component stars are so close together they touch. Varying by a quarter of a magnitude around an average apparent magnitude of 4.83 every seven hours, the star system lies 98 light years away. Its spectral type is F4VFe-0.8+. At the southern end of the crown asterism are the stars Eta¹ and Eta² Coronae Australis, which form an optical double. Of magnitude 5.1 and 5.5, they are separable with the naked eye and are both white. Kappa Coronae Australis is an easily resolved optical double—the components are of apparent magnitudes 6.3 and 5.6 and are about 1000 and 150 light years away respectively. They appear at an angle of 359 degrees, separated by 21.6 arcseconds. Kappa² is actually the brighter of the pair and is more bluish white, with a spectral type of B9V, while Kappa¹ is of spectral type A0III. Lying 202 light years away, Lambda Coronae Australis is a double splittable in small telescopes. The primary is a white star of spectral type A2Vn and magnitude of 5.1, while the companion star has a magnitude of 9.7. The two components are separated by 29.2 arcseconds at an angle of 214 degrees. Zeta Coronae Australis is a rapidly rotating main sequence star with an apparent magnitude of 4.8, 221.7 light years from Earth. The star has blurred lines in its hydrogen spectrum due to its rotation. Its spectral type is B9V. Theta Coronae Australis lies further to the west, a yellow giant of spectral type G8III and apparent magnitude 4.62. Corona Australis harbours RX J1856.5-3754, an isolated neutron star that is thought to lie 140 (±40) parsecs, or 460 (±130) light years, away, with a diameter of 14 km. It was once suspected to be a strange star, but this has been discounted. Corona Australis Molecular Cloud The Corona Australis Molecular Cloud is a dark molecular cloud just north of Beta Coronae Australis. Illuminated by a number of embedded reflection nebulae the cloud fans out from Epsilon Coronae Australis eastward along the constellation border with Sagittarius. It contains , Herbig–Haro objects (protostars) and some very young stars, being one of the closest star-forming regions, 430 light years (130 parsecs) to the Solar System, at the surface of the Local Bubble. The first nebulae of the cloud were recorded in 1865 by Johann Friedrich Julius Schmidt. Between Epsilon and Gamma Coronae Australis the cloud consists of the particular dark nebula and star forming region Bernes 157. It is 55 by 18 arcminutes wide and possesses several stars around magnitude 13. These stars are dimmed by up to 8 magnitudes because of the obscuring dust clouds. At the center of the active star-forming region lies the Coronet cluster (also called R CrA Cluster), which is used in studying star and protoplanetary disk formation. R Coronae Australis (R CrA) is an irregular variable star ranging from magnitudes 9.7 to 13.9. Blue-white, it is of spectral type B5IIIpe. A very young star, it is still accumulating interstellar material. It is obscured by, and illuminates, the surrounding nebula, NGC 6729, which brightens and darkens with it. The nebula is often compared to a comet for its appearance in a telescope, as its length is five times its width. Other stars of the cluster include S Coronae Australis, a G-class dwarf and T Tauri star. Nearby north, another young variable star, TY Coronae Australis, illuminates another nebula: reflection nebula NGC 6726/NGC 6727. TY Coronae Australis ranges irregularly between magnitudes 8.7 and 12.4, and the brightness of the nebula varies with it. Blue-white, it is of spectral type B8e. The largest young stars in the region, R, S, T, TY and VV Coronae Australis, are all ejecting jets of material which cause surrounding dust and gas to coalesce and form Herbig–Haro objects, many of which have been identified nearby. Not part of it is the globular cluster known as NGC 6723, which can be seen adjacent to the nebulosity in the neighbouring constellation of Sagittarius, but is much much further away. Deep sky objects IC 1297 is a planetary nebula of apparent magnitude 10.7, which appears as a green-hued roundish object in higher-powered amateur instruments. The nebula surrounds the variable star RU Coronae Australis, which has an average apparent magnitude of 12.9 and is a WC class Wolf–Rayet star. IC 1297 is small, at only 7 arcseconds in diameter; it has been described as "a square with rounded edges" in the eyepiece, elongated in the north–south direction. Descriptions of its color encompass blue, blue-tinged green, and green-tinged blue. Corona Australis' location near the Milky Way means that galaxies are uncommonly seen. NGC 6768 is a magnitude 11.2 object 35′ south of IC 1297. It is made up of two galaxies merging, one of which is an elongated elliptical galaxy of classification E4 and the other a lenticular galaxy of classification S0. IC 4808 is a galaxy of apparent magnitude 12.9 located on the border of Corona Australis with the neighbouring constellation of Telescopium and 3.9 degrees west-southwest of Beta Sagittarii. However, amateur telescopes will only show a suggestion of its spiral structure. It is 1.9 arcminutes by 0.8 arcminutes. The central area of the galaxy does appear brighter in an amateur instrument, which shows it to be tilted northeast–southwest. Southeast of Theta and southwest of Eta lies the open cluster ESO 281-SC24, which is composed of the yellow 9th magnitude star GSC 7914 178 1 and five 10th to 11th magnitude stars. Halfway between Theta Coronae Australis and Theta Scorpii is the dense globular cluster NGC 6541. Described as between magnitude 6.3 and magnitude 6.6, it is visible in binoculars and small telescopes. Around 22000 light years away, it is around 100 light years in diameter. It is estimated to be around 14 billion years old. NGC 6541 appears 13.1 arcminutes in diameter and is somewhat resolvable in large amateur instruments; a 12-inch telescope reveals approximately 100 stars but the core remains unresolved. Meteor showers The Corona Australids are a meteor shower that takes place between 14 and 18 March each year, peaking around 16 March. This meteor shower does not have a high peak hourly rate. In 1953 and 1956, observers noted a maximum of 6 meteors per hour and 4 meteors per hour respectively; in 1955 the shower was "barely resolved". However, in 1992, astronomers detected a peak rate of 45 meteors per hour. The Corona Australids' rate varies from year to year. At only six days, the shower's duration is particularly short, and its meteoroids are small; the stream is devoid of large meteoroids. The Corona Australids were first seen with the unaided eye in 1935 and first observed with radar in 1955. Corona Australid meteors have an entry velocity of 45 kilometers per second. In 2006, a shower originating near Beta Coronae Australis was designated as the Beta Coronae Australids. They appear in May, the same month as a nearby shower known as the May Microscopids, but the two showers have different trajectories and are unlikely to be related. History Corona Australis may have been recorded by ancient Mesopotamians in the MUL.APIN, as a constellation called MA.GUR ("The Bark"). However, this constellation, adjacent to SUHUR.MASH ("The Goat-Fish", modern Capricornus), may instead have been modern Epsilon Sagittarii. As a part of the southern sky, MA.GUR was one of the fifteen "stars of Ea". In the 3rd century BC, the Greek didactic poet Aratus wrote of, but did not name the constellation, instead calling the two crowns Στεφάνοι (Stephanoi). The Greek astronomer Ptolemy described the constellation in the 2nd century AD, though with the inclusion of Alpha Telescopii, since transferred to Telescopium. Ascribing 13 stars to the constellation, he named it Στεφάνος νοτιος (), "Southern Wreath", while other authors associated it with either Sagittarius (having fallen off his head) or Centaurus; with the former, it was called Corona Sagittarii. Similarly, the Romans called Corona Australis the "Golden Crown of Sagittarius". It was known as Parvum Coelum ("Canopy", "Little Sky") in the 5th century. The 18th-century French astronomer Jérôme Lalande gave it the names Sertum Australe ("Southern Garland") and Orbiculus Capitis, while German poet and author Philippus Caesius called it Corolla ("Little Crown") or Spira Australis ("Southern Coil"), and linked it with the Crown of Eternal Life from the New Testament. Seventeenth-century celestial cartographer Julius Schiller linked it to the Diadem of Solomon. Sometimes, Corona Australis was not the wreath of Sagittarius but arrows held in his hand. Corona Australis has been associated with the myth of Bacchus and Stimula. Jupiter had impregnated Stimula, causing Juno to become jealous. Juno convinced Stimula to ask Jupiter to appear in his full splendor, which the mortal woman could not handle, causing her to burn. After Bacchus, Stimula's unborn child, became an adult and the god of wine, he honored his deceased mother by placing a wreath in the sky. In Chinese astronomy, the stars of Corona Australis are located within the Black Tortoise of the North (北方玄武, Běi Fāng Xuán Wǔ). The constellation itself was known as ti'en pieh ("Heavenly Turtle") and during the Western Zhou period, marked the beginning of winter. However, precession over time has meant that the "Heavenly River" (Milky Way) became the more accurate marker to the ancient Chinese and hence supplanted the turtle in this role. Arabic names for Corona Australis include Al Ķubbah "the Tortoise", Al Ĥibā "the Tent" or Al Udḥā al Na'ām "the Ostrich Nest". It was later given the name Al Iklīl al Janūbiyyah, which the European authors Chilmead, Riccioli and Caesius transliterated as Alachil Elgenubi, Elkleil Elgenubi and Aladil Algenubi respectively. The ǀXam speaking San people of South Africa knew the constellation as ≠nabbe ta !nu "house of branches"—owned originally by the Dassie (rock hyrax), and the star pattern depicting people sitting in a semicircle around a fire. The indigenous Boorong people of northwestern Victoria saw it as Won, a boomerang thrown by Totyarguil (Altair). The Aranda people of Central Australia saw Corona Australis as a coolamon carrying a baby, which was accidentally dropped to earth by a group of sky-women dancing in the Milky Way. The impact of the coolamon created Gosses Bluff crater, 175 km west of Alice Springs. The Torres Strait Islanders saw Corona Australis as part of a larger constellation encompassing part of Sagittarius and the tip of Scorpius's tail; the Pleiades and Orion were also associated. This constellation was Tagai's canoe, crewed by the Pleiades, called the Usiam, and Orion, called the Seg. The myth of Tagai says that he was in charge of this canoe, but his crewmen consumed all of the supplies onboard without asking permission. Enraged, Tagai bound the Usiam with a rope and tied them to the side of the boat, then threw them overboard. Scorpius's tail represents a suckerfish, while Eta Sagittarii and Theta Coronae Australis mark the bottom of the canoe. On the island of Futuna, the figure of Corona Australis was called Tanuma and in the Tuamotus, it was called Na Kaua-ki-Tonga. See also Corona Australis (Chinese astronomy) Chamaeleon complex References Citations Sources Online sources SIMBAD External links The Deep Photographic Guide to the Constellations: Corona Australis Corona Australis photo Warburg Institute Iconographic Database (medieval and early modern images of Corona Australis) Constellations Constellations listed by Ptolemy Southern constellations
6607
https://en.wikipedia.org/wiki/Chaosium
Chaosium
Chaosium Inc. is a publisher of tabletop role-playing games established by Greg Stafford in 1975. Chaosium's major titles include Call of Cthulhu, based on the horror fiction stories of H. P. Lovecraft, RuneQuest Glorantha, Pendragon, based on Thomas Mallory's Le Morte d'Arthur, and 7th Sea, "swashbuckling and sorcery" set in a fantasy 17th century Europe. Many of Chaosium’s product lines are based upon literary sources. While Stafford himself has been described as "one of the most decorated game designers of all time" and "the grand shaman of gaming", multiple other notable game designers have written for Chaosium. These include David Conyers, Matthew Costello, Larry DiTillio, Paul Fricker, David A. Hargrave, Rob Heinsoo, Keith Herber, Jennell Jaquays, Katharine Kerr, Reiner Knizia, Charlie Krank, Robin Laws, Penelope Love, Mark Morrison, Steve Perrin, Sandy Petersen, Ken Rolston, Ken St. Andre, Jonathan Tweet, John Wick, and Lynn Willis, among others. History 1975–1980: Early years Greg Stafford founded "The Chaosium" in 1975, deriving the name partly from his home, which was near the Oakland Coliseum, combining "coliseum" with "chaos". His purpose was to publish his first board game White Bear and Red Moon (later renamed Dragon Pass), a board game set in his fantasy world of Glorantha. In 1978, Chaosium published Steve Perrin's roleplaying game RuneQuest, also set in Glorantha, following up with a second edition in 1980 and various supplements over the next six years. 1980s: Growth and licensing with Avalon Hill In 1980, the company officially incorporated as Chaosium Inc. That year, Stafford and Lynn Willis simplified the RuneQuest rules into the 16-page Basic Role-Playing (BRP). These simulationist, skill-based generic rules formed the basis of many of Chaosium's later "d100" RPGs, most notably Call of Cthulhu, first published in 1982. Chaosium entered into a licensing agreement with Avalon Hill in 1983 to produce a third edition of RuneQuest. Avalon Hill manufactured and marketed the game, while Chaosium was responsible for acquisitions, design, development, and layout. Ken Rolston managed the line as "Rune Czar". One of the first RPGs by a female lead designer was published by Chaosium: Kerie Campbell-Robson's 1986 release Hawkmoon. 1986 also saw the release of Ghostbusters with West End Games. Designed by Sandy Petersen, Lynn Willis, and Greg Stafford, it was the first RPG to use the dice pool mechanic. West End would also use the system as the basis of Star Wars: The Roleplaying Game and, eventually, the D6 System. Late 1990s–early 2010s: Financial struggle In 1996, it was prematurely reported that Chaosium had secured the rights to publish a collectible card game based on the video game Doom. In 1998, following the financial failure of the collectible card game Mythos, Greg Stafford resigned as Chaosium president and left the company, along with Sandy Petersen (although they both remained shareholders). Chaosium effectively split up into various successor companies, each maintaining its focus on a few of the company's products. Stafford took the rights to his game setting Glorantha, setting up the company Issaries, Inc. to continue publishing this line (later licensing it to Moon Design Publications, along with the game HeroQuest). Long-time employees and part-owners Charlie Krank and Lynn Willis remained at Chaosium as President and Editor-in-Chief respectively, continuing on with Call of Cthulhu as the main product line. Lynn Willis retired in 2008 due to poor health and died in 2013. Mid 2010s: The return of Stafford and Petersen Problems and delays fulfilling the Kickstarters for the 7th edition of Call of Cthulhu led Stafford and Petersen to return to active roles at Chaosium in June 2015. Charlie Krank subsequently left the company. Later that year at Gen Con 2015, Stafford and Petersen announced Moon Design Publications were now part of the Chaosium ownership, and the four principals of Moon Design (Rick Meints, Jeff Richard, Michael O'Brien, and Neil Robinson) had become the new Chaosium management team. Chaosium once again became the licensed publisher for RuneQuest, HeroQuest, and other products related to Glorantha and continued to publish the Call of Cthulhu line. Stafford served as chair of the company board and creative consultant until his death in October 2018. Since retiring from the board in 2019, Petersen has done occasional freelance work for the company, as did original RuneQuest creator Steve Perrin until his death in 2021. As part of its financial reorganization, the new management closed the company office and warehouse in Hayward, California, ending Chaosium's long association with the San Francisco Bay Area. The company is now based in Ann Arbor, Michigan and uses a fulfillment house model for distribution of product. Delivery of the core rewards of the Call of Cthulhu 7th Edition Kickstarter finally commenced in April 2016. The new edition went on to win nine of the ten awards it was nominated for at the Gen Con 2017 ENnie Awards. Late 2010s to Present: Expansion After the consolidation and reorganization of the mid-decade, the company was again poised to expand its offerings through a combination of acquisitions, new licenses, and distribution deals. Greg Stafford's King Arthur Pendragon and Prince Valiant roleplaying games returned to Chaosium ownership on December 11, 2018. On April 2, 2019, Chaosium acquired the rights to the 7th Sea product line (both Second Edition and Khitai Kickstarters) from John Wick, including back stock of books published so far. On November 30, 2019, Chaosium acquired the rights to produce a role-playing game based on Ben Aaronovitch's Rivers of London urban fantasy novels. In the spring of 2020, Chaosium took over distribution of the English translations of Spanish fantasy game Aquelarre and French Ice Age roleplaying game Würm, both of which had been successfully kickstarted by Nocturnal Media. In February 2021, they added Upwind, an original game kickstarted by Nocturnal Media and Biohazard Games, to that list. On August 20, 2021, Chaosium acquired the rights to Cthulhu Britannica and World War Cthulhu, formerly produced under license by Cubicle 7 until 2017. On October 26, 2021, Moon Design Publishing announced it was forming a partnership with Black Monk Games of Poland and a new company, The Chaosium Group, was being formed to manage both. Fiction Chaosium began publishing a line of non-game books (primarily fiction) in 1993. Many titles are themed around H. P. Lovecraft's Cthulhu Mythos and related topics, although the first work published was Greg Stafford's fantasy work King of Sartar, set in his mythic world Glorantha. Cassilda's Song, a 2015 anthology based on Robert W. Chambers's King in Yellow and written entirely by women, was nominated for two 2016 World Fantasy Awards. In May, 2017, Chaosium appointed award-winning author and editor James Lowder as executive editor of fiction. Lowder had previously served as a consultant for Chaosium, helping the company and freelancers resolve payment and contract problems with past fiction projects. Although not published by Chaosium, the ongoing Wild Cards series of superhero science fiction originated from a long-running Superworld campaign gamemastered by Game of Thrones author George R. R. Martin and his circle of fellow writers who played in his game. Magazines Three magazines have been published by Chaosium to promote its products: Wyrm's Footnotes ran for fourteen issues from 1976 to 1982. For the first ten issues, it was a source of supporting material for White Bear and Red Moon. In 1981, starting with Issue #11, it became the official RuneQuest magazine. The last edition published during its initial run was Issue #14, dated April 1982. The magazine was revived in 2012 by Moon Design Publications, continuing the issue numbering at 15, despite the 30-year hiatus. Issues 1 to 14 were republished in PDF format in 2019. Different Worlds. Forty-seven bimonthly issues from Different Worlds were published. Chaosium, from 1979 to 1985, published the first thirty-eight and Sleuth Publications, from 1985 to 1987, the final nine. Tadashi Ehara was the editor of the magazine during the periods concerned by both publishing houses. Starry Wisdom, a Lovecraft-themed magazine, three issues of which Chaosium published in 1997. Reception Chaosium won the 2017 Silver Ennie Award for "Fan’s Choice for Best Publisher". See also Chaosium is part of the Bits and Mortar alliance. References External links of Chaosium A 3rd-party Chaosium history guide 1975 establishments in California American companies established in 1975 Companies based in Ann Arbor, Michigan Design companies established in 1975 ENnies winners History of Hayward, California History of Oakland, California Publishing companies established in 1975 Publishing companies of the United States Role-playing game publishing companies Trading card companies
6760
https://en.wikipedia.org/wiki/Cryonics
Cryonics
Cryonics (from kryos meaning 'cold') is the low-temperature freezing (usually at ) and storage of human remains, with the speculative hope that resurrection may be possible in the future. Cryonics is regarded with skepticism within the mainstream scientific community. It is generally viewed as a pseudoscience, and its practice has been characterized as quackery. Cryonics procedures can begin only after the "patients" are clinically and legally dead. Cryonics procedures may begin within minutes of death, and use cryoprotectants to try to prevent ice formation during cryopreservation. It is, however, not possible for a corpse to be reanimated after undergoing vitrification, as this causes damage to the brain including its neural circuits. The first corpse to be frozen was that of James Bedford in 1967. As of 2014, about 250 bodies had been cryopreserved in the United States, and 1,500 people had made arrangements for cryopreservation of their corpses. Critics argue that economic reality means it is highly improbable that any cryonics corporation could continue in business long enough to take advantage of the claimed long-term benefits offered. Early attempts of cryonic preservations were performed in the 1960s and early 1970s which ended in failure with all but one of the companies going out of business, and their stored corpses thawed and disposed of. Conceptual basis Cryonicists argue that as long as brain structure remains intact, there is no fundamental barrier, given our current understanding of physical law, to recovering its information content. Cryonics proponents go further than the mainstream consensus in saying that the brain does not have to be continuously active to survive or retain memory. Cryonics controversially states that a human survives even within an inactive brain that has been badly damaged, provided that original encoding of memory and personality can, in theory, be adequately inferred and reconstituted from what structure remains. Cryonics uses temperatures below −130 °C, called cryopreservation, in an attempt to preserve enough brain information to permit the future revival of the cryopreserved person. Cryopreservation may be accomplished by freezing, freezing with cryoprotectant to reduce ice damage, or by vitrification to avoid ice damage. Even using the best methods, cryopreservation of whole bodies or brains is very damaging and irreversible with current technology. Cryonics advocates hold that in the future the use of some kind of presently-nonexistent nanotechnology may be able to help bring the dead back to life and treat the diseases which killed them. Mind uploading has also been proposed. Cryonics in practice Cryonics can be expensive. , the cost of preparing and storing corpses using cryonics ranged from US$28,000 to $200,000. When used at high concentrations, cryoprotectants can stop ice formation completely. Cooling and solidification without crystal formation is called vitrification. The first cryoprotectant solutions able to vitrify at very slow cooling rates while still being compatible with whole organ survival were developed in the late 1990s by cryobiologists Gregory Fahy and Brian Wowk for the purpose of banking transplantable organs. This has allowed animal brains to be vitrified, warmed back up, and examined for ice damage using light and electron microscopy. No ice crystal damage was found; cellular damage was due to dehydration and toxicity of the cryoprotectant solutions. Costs can include payment for medical personnel to be on call for death, vitrification, transportation in dry ice to a preservation facility, and payment into a trust fund intended to cover indefinite storage in liquid nitrogen and future revival costs. As of 2011, U.S. cryopreservation costs can range from $28,000 to $200,000, and are often financed via life insurance. KrioRus, which stores bodies communally in large dewars, charges $12,000 to $36,000 for the procedure. Some customers opt to have only their brain cryopreserved ("neuropreservation"), rather than their whole body. As of 2014, about 250 corpses have been cryogenically preserved in the U.S., and around 1,500 people have signed up to have their remains preserved. As of 2016, four facilities exist in the world to retain cryopreserved bodies: three in the U.S. and one in Russia. Considering the lifecycle of corporations, it is extremely unlikely that any cryonics company could continue to exist for sufficient time to take advantage even of the supposed benefits offered: historically, even the most robust corporations have only a one-in-a-thousand chance of surviving even one hundred years. Many cryonics companies have failed; , all but one of the pre-1973 batch had gone out of business, and their stored corpses have been defrosted and disposed of. Obstacles to success Preservation damage Cryopreservation has long been used by medical laboratories to maintain animal cells, human embryos, and even some organized tissues, for periods as long as three decades. Recovering large animals and organs from a frozen state is however not considered possible at the current level of scientific knowledge Large vitrified organs tend to develop fractures during cooling, a problem worsened by the large tissue masses and very low temperatures of cryonics. Without cryoprotectants, cell shrinkage and high salt concentrations during freezing usually prevent frozen cells from functioning again after thawing. Ice crystals can also disrupt connections between cells that are necessary for organs to function. In 2016, Robert L. McIntyre and Gregory Fahy at the cryobiology research company 21st Century Medicine, Inc. won the Small Animal Brain Preservation Prize of the Brain Preservation Foundation by demonstrating to the satisfaction of neuroscientist judges that a particular implementation of fixation and vitrification called aldehyde-stabilized cryopreservation could preserve a rabbit brain in "near perfect" condition at −135 °C, with the cell membranes, synapses, and intracellular structures intact in electron micrographs. Brain Preservation Foundation President, Ken Hayworth, said, "This result directly answers a main skeptical and scientific criticism against cryonics—that it does not provably preserve the delicate synaptic circuitry of the brain." However, the price paid for perfect preservation, as seen by microscopy, was tying up all protein molecules with chemical crosslinks, eliminating biological viability. Some cryonics organizations use vitrification without a chemical fixation step, sacrificing some structural preservation quality for less damage at the molecular level. Some scientists, like João Pedro Magalhães, have questioned whether using a deadly chemical for fixation eliminates the possibility of biological revival, making chemical fixation unsuitable for cryonics. Outside of cryonics firms and cryonics-linked interest groups, many scientists show strong skepticism toward cryonics methods. Cryobiologist Dayong Gao states that "we simply don't know if (subjects have) been damaged to the point where they've 'died' during vitrification because the subjects are now inside liquid nitrogen canisters." Biochemist Ken Storey argues (based on experience with organ transplants), that "even if you only wanted to preserve the brain, it has dozens of different areas, which would need to be cryopreserved using different protocols." Revival Revival would require repairing damage from lack of oxygen, cryoprotectant toxicity, thermal stress (fracturing) and freezing in tissues that do not successfully vitrify, finally followed by reversing the cause of death. In many cases, extensive tissue regeneration would be necessary. This revival technology remains speculative and does not currently exist. Legal issues Historically, a person had little control regarding how their body was treated after death as religion held jurisdiction over the ultimate fate of their body. However, secular courts began to exercise jurisdiction over the body and use discretion in carrying out of the wishes of the deceased person. Most countries legally treat preserved individuals as deceased persons because of laws that forbid vitrifying someone who is medically alive. In France, cryonics is not considered a legal mode of body disposal; only burial, cremation, and formal donation to science are allowed. However, bodies may legally be shipped to other countries for cryonic freezing. As of 2015, the Canadian province of British Columbia prohibits the sale of arrangements for body preservation based on cryonics. In Russia, cryonics falls outside both the medical industry and the funeral services industry, making it easier in Russia than in the U.S. to get hospitals and morgues to release cryonics candidates. In London in 2016, the English High Court ruled in favor of a mother's right to seek cryopreservation of her terminally ill 14-year-old daughter, as the girl wanted, contrary to the father's wishes. The decision was made on the basis that the case represented a conventional dispute over the disposal of the girl's body, although the judge urged ministers to seek "proper regulation" for the future of cryonic preservation following concerns raised by the hospital about the competence and professionalism of the team that conducted the preservation procedures. In Alcor Life Extension Foundation v. Richardson, the Iowa Court of Appeals ordered for the disinterment of Richardson, who was buried against his wishes, for cryopreservation. A detailed legal examination by Jochen Taupitz concludes that cryonic storage is legal in Germany for an indefinite period of time. Ethics In 2009, writing in Bioethics, David Shaw examines the ethical status of cryonics. The arguments against it include changing the concept of death, the expense of preservation and revival, lack of scientific advancement to permit revival, temptation to use premature euthanasia, and failure due to catastrophe. Arguments in favor of cryonics include the potential benefit to society, the prospect of immortality, and the benefits associated with avoiding death. Shaw explores the expense and the potential payoff, and applies an adapted version of Pascal's Wager to the question. In 2016, Charles Tandy wrote in favor of cryonics, arguing that honoring someone's last wishes is seen as a benevolent duty in American and many other cultures. History Cryopreservation was applied to human cells beginning in 1954 with frozen sperm, which was thawed and used to inseminate three women. The freezing of humans was first scientifically proposed by Michigan professor Robert Ettinger when he wrote The Prospect of Immortality (1962). In April 1966, the first human body was frozen—though it had been embalmed for two months—by being placed in liquid nitrogen and stored at just above freezing. The middle-aged woman from Los Angeles, whose name is unknown, was soon thawed out and buried by relatives. The first body to be cryopreserved and then frozen with the hope of future revival was that of James Bedford, claimed by Alcor's Mike Darwin to have occurred within around two hours of his death from cardiorespiratory arrest (secondary to metastasized kidney cancer) on January 12, 1967. Bedford's corpse is the only one frozen before 1974 still preserved today. In 1976, Ettinger founded the Cryonics Institute; his corpse was cryopreserved in 2011. Robert Nelson, "a former TV repairman with no scientific background" who led the Cryonics Society of California, was sued in 1981 for allowing nine bodies to thaw and decompose in the 1970s; in his defense, he claimed that the Cryonics Society had run out of money. This led to the lowered reputation of cryonics in the U.S. In 2018, a Y-Combinator startup called Nectome was recognized for developing a method of preserving brains with chemicals rather than by freezing. The method is fatal, performed as euthanasia under general anesthesia, but the hope is that future technology would allow the brain to be physically scanned into a computer simulation, neuron by neuron. Demographics According to The New York Times, cryonicists are predominantly non-religious white males, outnumbering women by about three to one. According to The Guardian, as of 2008, while most cryonicists used to be young, male, and "geeky", recent demographics have shifted slightly towards whole families. In 2015, Du Hong, a 61-year-old female writer of children's literature, became the first known Chinese national to have her head cryopreserved. Reception Cryonics is generally regarded as a fringe pseudoscience. The Society for Cryobiology rejected members who practiced cryonics, and issued a public statement saying that cryonics is "not science", and that it is a "personal choice" how people want to have their dead bodies disposed of. Russian company KrioRus is the first non-US vendor of cryonics services. Yevgeny Alexandrov, chair of the Russian Academy of Sciences commission against pseudoscience, said there was "no scientific basis" for cryonics, and that the company's offering was based on "unfounded speculation". Scientists have expressed skepticism about cryonics in media sources, and the Norwegian philosopher Ole Martin Moen has written that the topic receives a "minuscule" amount of attention from academia. While some neuroscientists contend that all the subtleties of a human mind are contained in its anatomical structure, few neuroscientists will comment directly upon the topic of cryonics due to its speculative nature. Individuals who intend to be frozen are often "looked at as a bunch of kooks". Cryobiologist Kenneth B. Storey said in 2004 that cryonics is impossible and will never be possible, as cryonics proponents are proposing to "over-turn the laws of physics, chemistry, and molecular science". Neurobiologist Michael Hendricks has said that "Reanimation or simulation is an abjectly false hope that is beyond the promise of technology and is certainly impossible with the frozen, dead tissue offered by the 'cryonics' industry". Anthropologist Simon Dein write that cryonics is a typical pseudoscience because of its lack of falsifiability and testability. In Dein's view cryonics is not science, but religion: it places faith in non-existent technology and promises to overcome death itself. William T. Jarvis has written that "Cryonics might be a suitable subject for scientific research, but marketing an unproven method to the public is quackery". According to cryonicist Aschwin de Wolf and others, cryonics can often produce intense hostility from spouses who are not cryonicists. James Hughes, the executive director of the pro-life-extension Institute for Ethics and Emerging Technologies, chooses not to personally sign up for cryonics, calling it a worthy experiment but stating laconically that "I value my relationship with my wife." Cryobiologist Dayong Gao states that "People can always have hope that things will change in the future, but there is no scientific foundation supporting cryonics at this time." While it is universally agreed that "personal identity" is uninterrupted when brain activity temporarily ceases during incidents of accidental drowning (where people have been restored to normal functioning after being completely submerged in cold water for up to 66 minutes), one argument against cryonics is that a centuries-long absence from life might interrupt the conception of personal identity, such that the revived person would "not be themself". Maastricht University bioethicist David Shaw raises the argument that there would be no point in being revived in the far future if one's friends and families are dead, leaving them all alone; he notes, however, that family and friends can also be frozen, that there is "nothing to prevent the thawed-out freezee from making new friends", and that a lonely existence may be preferable to no existence at all for the revived. In fiction Suspended animation is a popular subject in science fiction and fantasy settings. It is often the means by which a character is transported into the future. A survey in Germany found that about half of the respondents were familiar with cryonics, and about half of those familiar with cryonics had learned of the subject from films or television. In popular culture The town of Nederland, Colorado, hosts an annual Frozen Dead Guy Days festival to commemorate a substandard attempt at cryopreservation. Notable people Corpses subjected to the cryonics process include those of baseball players Ted Williams and son John Henry Williams (in 2002 and 2004, respectively), engineer and doctor L. Stephen Coles (in 2014), economist and entrepreneur Phil Salin, and software engineer Hal Finney (in 2014). People known to have arranged for cryonics upon death include PayPal founders Luke Nosek and Peter Thiel, Oxford transhumanists Nick Bostrom and Anders Sandberg, and transhumanist philosopher David Pearce. Larry King previously arranged for cryonics, but according to Inside Edition, later changed his mind. Disgraced financier Jeffrey Epstein wanted to have his head and penis frozen after death so that he could "seed the human race with his DNA." The corpses of some are mistakenly believed to have undergone cryonics – for instance, the urban legend suggesting Walt Disney's corpse was cryopreserved is false; it was cremated and interred at Forest Lawn Memorial Park Cemetery. Robert A. Heinlein, who wrote enthusiastically of the concept in The Door into Summer (serialized in 1956), was cremated and had his ashes distributed over the Pacific Ocean. Timothy Leary was a long-time cryonics advocate and signed up with a major cryonics provider, but he changed his mind shortly before his death and was not cryopreserved. See also Brain in a vat Cryptobiosis Extropianism Hibernation Life extension Supercooling References Footnotes Citations Further reading External links Cooling technology Death customs Fictional technology Health fraud Life extension Pseudoscience Scientific speculation
6920
https://en.wikipedia.org/wiki/Column
Column
A column or pillar in architecture and structural engineering is a structural element that transmits, through compression, the weight of the structure above to other structural elements below. In other words, a column is a compression member. The term column applies especially to a large round support (the shaft of the column) with a capital and a base or pedestal, which is made of stone, or appearing to be so. A small wooden or metal support is typically called a post. Supports with a rectangular or other non-round section are usually called piers. For the purpose of wind or earthquake engineering, columns may be designed to resist lateral forces. Other compression members are often termed "columns" because of the similar stress conditions. Columns are frequently used to support beams or arches on which the upper parts of walls or ceilings rest. In architecture, "column" refers to such a structural element that also has certain proportional and decorative features. A column might also be a decorative element not needed for structural purposes; many columns are engaged, that is to say form part of a wall. A long sequence of columns joined by an entablature is known as a colonnade. History Antiquity All significant Iron Age civilizations of the Near East and Mediterranean made some use of columns. Israelite The Jews of Eretz Yisrael used columns beginning with the Tabernacle following the Exodus from Egypt and in the First (Solomon's) Temple. Additionally, the Second (Herod's) Temple included columns, albeit with Roman and Greek influences. Egyptian In ancient Egyptian architecture as early as 2600 BC, the architect Imhotep made use of stone columns whose surface was carved to reflect the organic form of bundled reeds, like papyrus, lotus and palm. In later Egyptian architecture faceted cylinders were also common. Their form is thought to derive from archaic reed-built shrines. Carved from stone, the columns were highly decorated with carved and painted hieroglyphs, texts, ritual imagery and natural motifs. Egyptian columns are famously present in the Great Hypostyle Hall of Karnak (circa 1224 BC), where 134 columns are lined up in sixteen rows, with some columns reaching heights of 24 metres. One of the most important type are the papyriform columns. The origin of these columns goes back to the 5th Dynasty. They are composed of lotus (papyrus) stems which are drawn together into a bundle decorated with bands: the capital, instead of opening out into the shape of a bellflower, swells out and then narrows again like a flower in bud. The base, which tapers to take the shape of a half-sphere like the stem of the lotus, has a continuously recurring decoration of stipules. Greek and Roman The Minoans used whole tree-trunks, usually turned upside down in order to prevent re-growth, stood on a base set in the stylobate (floor base) and topped by a simple round capital. These were then painted as in the most famous Minoan palace of Knossos. The Minoans employed columns to create large open-plan spaces, light-wells and as a focal point for religious rituals. These traditions were continued by the later Mycenaean civilization, particularly in the megaron or hall at the heart of their palaces. The importance of columns and their reference to palaces and therefore authority is evidenced in their use in heraldic motifs such as the famous lion-gate of Mycenae where two lions stand each side of a column. Being made of wood these early columns have not survived, but their stone bases have and through these we may see their use and arrangement in these palace buildings. The Egyptians, Persians and other civilizations mostly used columns for the practical purpose of holding up the roof inside a building, preferring outside walls to be decorated with reliefs or painting, but the Ancient Greeks, followed by the Romans, loved to use them on the outside as well, and the extensive use of columns on the interior and exterior of buildings is one of the most characteristic features of classical architecture, in buildings like the Parthenon. The Greeks developed the classical orders of architecture, which are most easily distinguished by the form of the column and its various elements. Their Doric, Ionic, and Corinthian orders were expanded by the Romans to include the Tuscan and Composite orders. Persian Some of the most elaborate columns in the ancient world were those of the Persians, especially the massive stone columns erected in Persepolis. They included double-bull structures in their capitals. The Hall of Hundred Columns at Persepolis, measuring 70 × 70 metres, was built by the Achaemenid king Darius I (524–486 BC). Many of the ancient Persian columns are standing, some being more than 30 metres tall. Tall columns with bull's head capitals were used for porticoes and to support the roofs of the hypostylehall, partly inspired by the ancient Egyptian precedent. Since the columns carried timber beams rather than stone, they could be taller, slimmer and more widely spaced than Egyptian ones. Middle Ages Columns, or at least large structural exterior ones, became much less significant in the architecture of the Middle Ages. The classical forms were abandoned in both Byzantine and Romanesque architecture in favour of more flexible forms, with capitals often using various types of foliage decoration, and in the West scenes with figures carved in relief. During the Romanesque period, builders continued to reuse and imitate ancient Roman columns wherever possible; where new, the emphasis was on elegance and beauty, as illustrated by twisted columns. Often they were decorated with mosaics. Renaissance and later styles Renaissance architecture was keen to revive the classical vocabulary and styles, and the informed use and variation of the classical orders remained fundamental to the training of architects throughout Baroque, Rococo and Neo-classical architecture. Structure Early columns were constructed of stone, some out of a single piece of stone. Monolithic columns are among the heaviest stones used in architecture. Other stone columns are created out of multiple sections of stone, mortared or dry-fit together. In many classical sites, sectioned columns were carved with a centre hole or depression so that they could be pegged together, using stone or metal pins. The design of most classical columns incorporates entasis (the inclusion of a slight outward curve in the sides) plus a reduction in diameter along the height of the column, so that the top is as little as 83% of the bottom diameter. This reduction mimics the parallax effects which the eye expects to see, and tends to make columns look taller and straighter than they are while entasis adds to that effect. There are flutes and fillets that run up the shaft of columns. The flute is the part of the column that is indented in with a semi circular shape. The fillet of the column is the part between each of the flutes on the Ionic order columns. The flute width changes on all tapered columns as it goes up the shaft and stays the same on all non tapered columns. This was done to the columns to add visual interest to them. The Ionic and the Corinthian are the only orders that have fillets and flutes. The Doric style has flutes but not fillets. Doric flutes are connected at a sharp point where the fillets are located on Ionic and Corinthian order columns. Nomenclature Most classical columns arise from a basis, or base, that rests on the stylobate, or foundation, except for those of the Doric order, which usually rest directly on the stylobate. The basis may consist of several elements, beginning with a wide, square slab known as a plinth. The simplest bases consist of the plinth alone, sometimes separated from the column by a convex circular cushion known as a torus. More elaborate bases include two toruses, separated by a concave section or channel known as a scotia or trochilus. Scotiae could also occur in pairs, separated by a convex section called an astragal, or bead, narrower than a torus. Sometimes these sections were accompanied by still narrower convex sections, known as annulets or fillets. At the top of the shaft is a capital, upon which the roof or other architectural elements rest. In the case of Doric columns, the capital usually consists of a round, tapering cushion, or echinus, supporting a square slab, known as an abax or abacus. Ionic capitals feature a pair of volutes, or scrolls, while Corinthian capitals are decorated with reliefs in the form of acanthus leaves. Either type of capital could be accompanied by the same moldings as the base. In the case of free-standing columns, the decorative elements atop the shaft are known as a finial. Modern columns may be constructed out of steel, poured or precast concrete, or brick, left bare or clad in an architectural covering, or veneer. Used to support an arch, an impost, or pier, is the topmost member of a column. The bottom-most part of the arch, called the springing, rests on the impost. Equilibrium, instability, and loads As the axial load on a perfectly straight slender column with elastic material properties is increased in magnitude, this ideal column passes through three states: stable equilibrium, neutral equilibrium, and instability. The straight column under load is in stable equilibrium if a lateral force, applied between the two ends of the column, produces a small lateral deflection which disappears and the column returns to its straight form when the lateral force is removed. If the column load is gradually increased, a condition is reached in which the straight form of equilibrium becomes so-called neutral equilibrium, and a small lateral force will produce a deflection that does not disappear and the column remains in this slightly bent form when the lateral force is removed. The load at which neutral equilibrium of a column is reached is called the critical or buckling load. The state of instability is reached when a slight increase of the column load causes uncontrollably growing lateral deflections leading to complete collapse. For an axially loaded straight column with any end support conditions, the equation of static equilibrium, in the form of a differential equation, can be solved for the deflected shape and critical load of the column. With hinged, fixed or free end support conditions the deflected shape in neutral equilibrium of an initially straight column with uniform cross section throughout its length always follows a partial or composite sinusoidal curve shape, and the critical load is given by where E = elastic modulus of the material, Imin = the minimal moment of inertia of the cross section, and L = actual length of the column between its two end supports. A variant of (1) is given by where r = radius of gyration of column cross-section which is equal to the square root of (I/A), K = ratio of the longest half sine wave to the actual column length, Et = tangent modulus at the stress Fcr, and KL = effective length (length of an equivalent hinged-hinged column). From Equation (2) it can be noted that the buckling strength of a column is inversely proportional to the square of its length. When the critical stress, Fcr (Fcr =Pcr/A, where A = cross-sectional area of the column), is greater than the proportional limit of the material, the column is experiencing inelastic buckling. Since at this stress the slope of the material's stress-strain curve, Et (called the tangent modulus), is smaller than that below the proportional limit, the critical load at inelastic buckling is reduced. More complex formulas and procedures apply for such cases, but in its simplest form the critical buckling load formula is given as Equation (3), A column with a cross section that lacks symmetry may suffer torsional buckling (sudden twisting) before, or in combination with, lateral buckling. The presence of the twisting deformations renders both theoretical analyses and practical designs rather complex. Eccentricity of the load, or imperfections such as initial crookedness, decreases column strength. If the axial load on the column is not concentric, that is, its line of action is not precisely coincident with the centroidal axis of the column, the column is characterized as eccentrically loaded. The eccentricity of the load, or an initial curvature, subjects the column to immediate bending. The increased stresses due to the combined axial-plus-flexural stresses result in a reduced load-carrying ability. Column elements are considered to be massive if their smallest side dimension is equal to or more than 400 mm. Massive columns have the ability to increase in carrying strength over long time periods (even during periods of heavy load). Taking into account the fact, that possible structural loads may increase over time as well (and also the threat of progressive failure), massive columns have an advantage compared to non-massive ones. Extensions When a column is too long to be built or transported in one piece, it has to be extended or spliced at the construction site. A reinforced concrete column is extended by having the steel reinforcing bars protrude a few inches or feet above the top of the concrete, then placing the next level of reinforcing bars to overlap, and pouring the concrete of the next level. A steel column is extended by welding or bolting splice plates on the flanges and webs or walls of the columns to provide a few inches or feet of load transfer from the upper to the lower column section. A timber column is usually extended by the use of a steel tube or wrapped-around sheet-metal plate bolted onto the two connecting timber sections. Foundations A column that carries the load down to a foundation must have means to transfer the load without overstressing the foundation material. Reinforced concrete and masonry columns are generally built directly on top of concrete foundations. When seated on a concrete foundation, a steel column must have a base plate to spread the load over a larger area, and thereby reduce the bearing pressure. The base plate is a thick, rectangular steel plate usually welded to the bottom end of the column. Orders The Roman author Vitruvius, relying on the writings (now lost) of Greek authors, tells us that the ancient Greeks believed that their Doric order developed from techniques for building in wood. The earlier smoothed tree-trunk was replaced by a stone cylinder. Doric order The Doric order is the oldest and simplest of the classical orders. It is composed of a vertical cylinder that is wider at the bottom. It generally has neither a base nor a detailed capital. It is instead often topped with an inverted frustum of a shallow cone or a cylindrical band of carvings. It is often referred to as the masculine order because it is represented in the bottom level of the Colosseum and the Parthenon, and was therefore considered to be able to hold more weight. The height-to-thickness ratio is about 8:1. The shaft of a Doric Column is almost always fluted. The Greek Doric, developed in the western Dorian region of Greece, is the heaviest and most massive of the orders. It rises from the stylobate without any base; it is from four to six times as tall as its diameter; it has twenty broad flutes; the capital consists simply of a banded necking swelling out into a smooth echinus, which carries a flat square abacus; the Doric entablature is also the heaviest, being about one-fourth the height column. The Greek Doric order was not used after c. 100 B.C. until its “rediscovery” in the mid-eighteenth century. Tuscan order The Tuscan order, also known as Roman Doric, is also a simple design, the base and capital both being series of cylindrical disks of alternating diameter. The shaft is almost never fluted. The proportions vary, but are generally similar to Doric columns. Height to width ratio is about 7:1. Ionic order The Ionic column is considerably more complex than the Doric or Tuscan. It usually has a base and the shaft is often fluted (it has grooves carved up its length). The capital features a volute, an ornament shaped like a scroll, at the four corners. The height-to-thickness ratio is around 9:1. Due to the more refined proportions and scroll capitals, the Ionic column is sometimes associated with academic buildings. Ionic style columns were used on the second level of the Colosseum. Corinthian order The Corinthian order is named for the Greek city-state of Corinth, to which it was connected in the period. However, according to the architectural historian Vitruvius, the column was created by the sculptor Callimachus, probably an Athenian, who drew acanthus leaves growing around a votive basket. In fact, the oldest known Corinthian capital was found in Bassae, dated at 427 BC. It is sometimes called the feminine order because it is on the top level of the Colosseum and holding up the least weight, and also has the slenderest ratio of thickness to height. Height to width ratio is about 10:1. Composite order The Composite order draws its name from the capital being a composite of the Ionic and Corinthian capitals. The acanthus of the Corinthian column already has a scroll-like element, so the distinction is sometimes subtle. Generally the Composite is similar to the Corinthian in proportion and employment, often in the upper tiers of colonnades. Height to width ratio is about 11:1 or 12:1. Solomonic A Solomonic column, sometimes called "barley sugar", begins on a base and ends in a capital, which may be of any order, but the shaft twists in a tight spiral, producing a dramatic, serpentine effect of movement. Solomonic columns were developed in the ancient world, but remained rare there. A famous marble set, probably 2nd century, was brought to Old St. Peter's Basilica by Constantine I, and placed round the saint's shrine, and was thus familiar throughout the Middle Ages, by which time they were thought to have been removed from the Temple of Jerusalem. The style was used in bronze by Bernini for his spectacular St. Peter's baldachin, actually a ciborium (which displaced Constantine's columns), and thereafter became very popular with Baroque and Rococo church architects, above all in Latin America, where they were very often used, especially on a small scale, as they are easy to produce in wood by turning on a lathe (hence also the style's popularity for spindles on furniture and stairs). Caryatid A Caryatid is a sculpted female figure serving as an architectural support taking the place of a column or a pillar supporting an entablature on her head. The Greek term literally means "maidens of Karyai", an ancient town of Peloponnese. Engaged columns In architecture, an engaged column is a column embedded in a wall and partly projecting from the surface of the wall, sometimes defined as semi or three-quarter detached. Engaged columns are rarely found in classical Greek architecture, and then only in exceptional cases, but in Roman architecture they exist in abundance, most commonly embedded in the cella walls of pseudoperipteral buildings. Pillar tombs Pillar tombs are monumental graves, which typically feature a single, prominent pillar or column, often made of stone. A number of world cultures incorporated pillars into tomb structures. In the ancient Greek colony of Lycia in Anatolia, one of these edifices is located at the tomb of Xanthos. In the town of Hannassa in southern Somalia, ruins of houses with archways and courtyards have also been found along with other pillar tombs, including a rare octagonal tomb. Gallery See also Columnar jointing (geology) Huabiao Linga Lingodbhava Marian and Holy Trinity columns Our Lady of the Pillar Post (structural) Pylon (architecture) Spur (architecture) Structural engineering References Chisholm, Hugh, ed. (1911). "Engaged Column". Encyclopædia Britannica. 9 (11th ed.). Cambridge University Press. pp. 404–405. Stierlin, Henri The Roman Empire: From the Etruscans to the Decline of the Roman Empire, TASCHEN, 2002 Alderman, Liz (7 July 2014). "Acropolis Maidens Glow Anew". The New York Times. Retrieved 9 July 2014. Stokstad, Marilyn; Cothren, Michael (2014). Art History (Volume 1 ed.). New Jersey: Pearson Education, Inc. p. 110. External links Architectural elements Structural system Earthquake engineering
7080
https://en.wikipedia.org/wiki/Christian%20Doppler
Christian Doppler
Christian Andreas Doppler ( (); 29 November 1803 – 17 March 1853) was an Austrian mathematician and physicist. He formulated the principle – now known as the Doppler effect – that the observed frequency of a wave depends on the relative speed of the source and the observer. Biography Early life and education Doppler was born in Salzburg (today Austria) in 1803. Doppler was the second son born to Johann Evangelist Doppler and Theresia Seeleuthner (Doppler). Doppler's father, Johann Doppler, was a third-generation stone mason in Salzburg. As a young boy, Doppler showed promise for his family's trade. However, due to his weak health, Doppler's father encouraged him instead to pursue a career in business. Doppler started elementary education at the age of 13. After completion, he moved on to secondary education at a school in Linz. Doppler's proficiency in mathematics was discovered by Sion Stampfer, a mathematician in Salzburg. Upon the recommendation of Sion Stampfer, Doppler took a break from high school to attend the Polytechnic Institute in Vienna in 1822. Doppler returned to Salzburg in 1825 to finish his secondary education. After completing high school, Doppler studied philosophy in Salzburg and mathematics and physics at the University of Vienna and Imperial–Royal Polytechnic Institute (now TU Wien). In 1829, he was chosen for an assistant position to Professor Adam Von Burg at the Polytechnic Institute of Vienna, where he continued his studies. In 1835, he decided to immigrate to the United States to pursue a position in academia. Before departing for the United States, Doppler was offered a teaching position at a state-operated high school in Prague, which convinced him to stay in Europe. Shortly after, in 1837 he was appointed as an associate professor of math and geometry at the Prague Polytechnic Institute (now Czech Technical University in Prague). He received a full professorship position in 1841. Family In 1836, Doppler married Mathilde Sturm, the daughter of goldsmith Franz Sturm. Doppler and Mathilde had five children together. Their first child was Mathilde Doppler who was born in 1837. Doppler's second child, Ludwig Doppler was born in 1838. Two years later, in 1840 Adolf Doppler was born. Doppler's fourth child, Bertha Doppler was born in 1843. Their last child Hermann was born in 1845. Development of the Doppler Effect In 1842, at the age of 38, Doppler gave a lecture to the Royal Bohemian Society of Sciences and subsequently published Über das farbige Licht der Doppelsterne und einiger anderer Gestirne des Himmels ("On the coloured light of the binary stars and some other stars of the heavens"). In this work, Doppler postulated his principle (later named the Doppler effect) that the observed frequency of a wave depends on the relative speed of the source and the observer, and he later tried to use this concept to explain the visible colours of binary stars (this hypothesis was later proven wrong). Doppler also incorrectly believed that if a star were to exceed 136,000 kilometers per second in radial velocity, then it would not be visible to the human eye. Later life Doppler continued working as a professor at the Prague Polytechnic, publishing over 50 articles on mathematics, physics and astronomy, but in 1847 he left Prague for the professorship of mathematics, physics, and mechanics at the Academy of Mines and Forests (its successor is the University of Miskolc) in Selmecbánya (then Kingdom of Hungary, now Banská Štiavnica Slovakia). Doppler's research was interrupted by the Hungarian Revolution of 1848. In 1849, he fled to Vienna and in 1850 was appointed head of the Institute for Experimental Physics at the University of Vienna. While there, Doppler, along with Franz Unger, influenced the development of young Gregor Mendel, the founding father of genetics, who was a student at the University of Vienna from 1851 to 1853. Death Doppler died on 17 March 1853 at age 49 from a pulmonary disease in Venice (at that time part of the Austrian Empire). His tomb is in the cemetery of the Venetian island of San Michele. Full name Some confusion exists about Doppler's full name. Doppler referred to himself as Christian Doppler. The records of his birth and baptism stated Christian Andreas Doppler. Doppler's middle name is shared by his great-great-grandfather Andreas Doppler. Forty years after Doppler's death the misnomer Johann Christian Doppler was introduced by the astronomer Julius Scheiner. Scheiner's mistake has since been copied by many. Works Christian Doppler (1803–1853). Wien: Böhlau, 1992. Bd. 1: 1. Teil: Helmuth Grössing (unter Mitarbeit von B. Reischl): Wissenschaft, Leben, Umwelt, Gesellschaft; 2. Teil: Karl Kadletz (unter Mitarbeit von Peter Schuster und Ildikó Cazan-Simányi) Quellenanhang. Bd. 2: 3. Teil: Peter Schuster: Das Werk. See also List of Austrian scientists List of Austrians List of minor planets named after people References Further reading Alec Eden: Christian Doppler: Leben und Werk. Salzburg: Landespressebureau, 1988. Hoffmann, Robert (2007). The Life of an (almost) Unknown Person. Christian Doppler's Youth in Salzburg and Vienna. In: Ewald Hiebl, Maurizio Musso (Eds.), Christian Doppler – Life and Work. Principle an Applications. Proceedings of the Commemorative Symposia in Salzburg, Salzburg, Prague, Vienna, Venice. Pöllauberg/Austria, Hainault/UK, Atascadero/US, pages 33 – 46. David Nolte (2020). The fall and rise of the Doppler effect. Physics Today, v. 73, pgs. 31 - 35. DOI: 10.1063/PT.3.4429 External links Christian Doppler Platform & Christian-Doppler-Fonds 1803 births 1853 deaths 19th-century Austrian mathematicians 19th-century Austrian physicists Austrian Roman Catholics People from Austria-Hungary Austrian expatriates in Hungary Scientists from Salzburg Burials at Isola di San Michele Czech Technical University in Prague alumni Academic staff of Czech Technical University in Prague Doppler effects
7257
https://en.wikipedia.org/wiki/Caste
Caste
A caste is a fixed social group into which an individual is born within a particular system of social stratification: a caste system. Within such a system, individuals are expected to: marry exclusively within the same caste (endogamy), follow lifestyles often linked to a particular occupation, hold a ritual status observed within a hierarchy, and interact with others based on cultural notions of exclusion, with certain castes considered as either more pure or more polluted than others. Its paradigmatic ethnographic example is the division of India's Hindu society into rigid social groups. Its roots lie in south Asia's ancient history and it still exists. However, the economic significance of the caste system in India has been declining as a result of urbanisation and affirmative action programs. A subject of much scholarship by sociologists and anthropologists, the Hindu caste system is sometimes used as an analogical basis for the study of caste-like social divisions existing outside Hinduism and India. The term "caste" is also applied to morphological groupings in eusocial insects such as ants, bees, and termites. Etymology The English word caste () derives from the Spanish and Portuguese , which, according to the John Minsheu's Spanish dictionary (1569), means "race, lineage, tribe or breed". When the Spanish colonised the New World, they used the word to mean a 'clan or lineage'. It was, however, the Portuguese who first employed in the primary modern sense of the English word 'caste' when they applied it to the thousands of endogamous, hereditary Indian social groups they encountered upon their arrival in India in 1498. The use of the spelling caste, with this latter meaning, is first attested in English in 1613. In the Latin American context, the term caste is sometimes used to describe the casta system of racial classification, based on whether a person was of pure European, Indigenous or African descent, or some mix thereof, with the different groups being placed in a racial hierarchy; however, despite the etymological connection between the Latin American casta system and South Asian caste systems (the former giving its name to the latter), it is controversial to what extent the two phenomena are really comparable. In South Asia India Modern India's caste system is based on the artificial modern superimposition of an old four-fold theoretical classification called the Varna on the natural social groupings called the Jāti. Varna conceptualised a society as consisting of four types of varnas, or categories: Brahmin, Kshatriya, Vaishya and Shudra, according to the nature of the work of its members. Varna was not an inherited category and the occupation determined the varna. However, a person's Jati is determined at birth and makes them take up that Jati's occupation; members could and did change their occupation based on personal strengths as well as economic, social and political factors. A 2016 study based on the DNA analysis of unrelated Indians determined that endogamous Jatis originated during the Gupta Empire. From 1901 onwards, for the purposes of the Decennial Census, the British colonial authorities arbitrarily and incorrectly forced all Jātis into the four Varna categories as described in ancient texts. Herbert Hope Risley, the Census Commissioner, noted that "The principle suggested as a basis was that of classification by social precedence as recognized by native public opinion at the present day, and manifesting itself in the facts that particular castes are supposed to be the modern representatives of one or other of the castes of the theoretical Indian system." Varna, as mentioned in ancient Hindu texts, describes society as divided into four categories: Brahmins (scholars and yajna priests), Kshatriyas (rulers and warriors), Vaishyas (farmers, merchants and artisans) and Shudras (workmen/service providers). The texts do not mention any hierarchy or a separate, untouchable category in Varna classifications. Scholars believe that the Varnas system was never truly operational in society and there is no evidence of it ever being a reality in Indian history. The practical division of the society had always been in terms of Jatis (birth groups), which are not based on any specific religious principle, but could vary from ethnic origins to occupations to geographic areas. The Jātis have been endogamous social groups without any fixed hierarchy but subject to vague notions of rank articulated over time based on lifestyle and social, political or economic status. Many of India's major empires and dynasties like the Mauryas, Shalivahanas, Chalukyas, Kakatiyas among many others, were founded by people who would have been classified as Shudras, under the Varnas system, as interpreted by the British rulers. It is well established that by the 9th century, kings from all the four Varnas, including Brahmins and Vaishyas, had occupied the highest seat in the monarchical system in Hindu India, contrary to the Varna theory. In many instances, as in Bengal, historically the kings and rulers had been called upon, when required, to mediate on the ranks of Jātis, which might number in thousands all over the subcontinent and vary by region. In practice, the jātis may or may not fit into the Varna classes and many prominent Jatis, for example the Jats and Yadavs, straddled two Varnas i.e. Kshatriyas and Vaishyas, and the Varna status of Jātis itself was subject to articulation over time. Starting with the 1901 Census of India led by colonial administrator Herbert Hope Risley, all the jātis were grouped under the theoretical varnas categories. According to political scientist Lloyd Rudolph, Risley believed that varna, however ancient, could be applied to all the modern castes found in India, and "[he] meant to identify and place several hundred million Indians within it." The terms varna (conceptual classification based on occupation) and jāti (groups) are two distinct concepts: while varna is a theoretical four-part division, jāti (community) refers to the thousands of actual endogamous social groups prevalent across the subcontinent. The classical authors scarcely speak of anything other than the varnas, as it provided a convenient shorthand; but a problem arises when colonial Indologists sometimes confuse the two. Upon independence from Britain, the Indian Constitution listed 1,108 Jatis across the country as Scheduled Castes in 1950, for positive discrimination. This constitution would also ban discrimination of the basis of the caste, though its practice in India remained intact. The Untouchable communities are sometimes called Scheduled Castes, Dalit or Harijan in contemporary literature. In 2001, Dalits were 16.2% of India's population. Most of the 15 million bonded child workers are from the lowest castes. Independent India has witnessed caste-related violence. In 2005, government recorded approximately 110,000 cases of reported violent acts, including rape and murder, against Dalits. For 2012, the government recorded 651 murders, 3,855 injuries, 1,576 rapes, 490 kidnappings, and 214 cases of arson. The socio-economic limitations of the caste system are reduced due to urbanisation and affirmative action. Nevertheless, the caste system still exists in endogamy and patrimony, and thrives in the politics of democracy, where caste provides ready made constituencies to politicians. The globalisation and economic opportunities from foreign businesses has influenced the growth of India's middle-class population. Some members of the Chhattisgarh Potter Caste Community (CPCC) are middle-class urban professionals and no longer potters unlike the remaining majority of traditional rural potter members. There is persistence of caste in Indian politics. Caste associations have evolved into caste-based political parties. Political parties and the state perceive caste as an important factor for mobilisation of people and policy development. Studies by Bhatt and Beteille have shown changes in status, openness, mobility in the social aspects of Indian society. As a result of modern socio-economic changes in the country, India is experiencing significant changes in the dynamics and the economics of its social sphere. While arranged marriages are still the most common practice in India, the internet has provided a network for younger Indians to take control of their relationships through the use of dating apps. This remains isolated to informal terms, as marriage is not often achieved through the use of these apps. Hypergamy is still a common practice in India and Hindu culture. Men are expected to marry within their caste, or one below, with no social repercussions. If a woman marries into a higher caste, then her children will take the status of their father. If she marries down, her family is reduced to the social status of their son in law. In this case, the women are bearers of the egalitarian principle of the marriage. There would be no benefit in marrying a higher caste if the terms of the marriage did not imply equality. However, men are systematically shielded from the negative implications of the agreement. Geographical factors also determine adherence to the caste system. Many Northern villages are more likely to participate in exogamous marriage, due to a lack of eligible suitors within the same caste. Women in North India have been found to be less likely to leave or divorce their husbands since they are of a relatively lower caste system, and have higher restrictions on their freedoms. On the other hand, Pahari women, of the northern mountains, have much more freedom to leave their husbands without stigma. This often leads to better husbandry as his actions are not protected by social expectations. Chiefly among the factors influencing the rise of exogamy is the rapid urbanisation in India experienced over the last century. It is well known that urban centers tend to be less reliant on agriculture and are more progressive as a whole. As India's cities boomed in population, the job market grew to keep pace. Prosperity and stability were now more easily attained by an individual, and the anxiety to marry quickly and effectively was reduced. Thus, younger, more progressive generations of urban Indians are less likely than ever to participate in the antiquated system of arranged endogamy. India has also implemented a form of Affirmative Action, locally known as "reservation groups". Quota system jobs, as well as placements in publicly funded colleges, hold spots for the 8% of India's minority, and underprivileged groups. As a result, in states such as Tamil Nadu or those in the north-east, where underprivileged populations predominate, over 80% of government jobs are set aside in quotas. In education, colleges lower the marks necessary for the Dalits to enter. Nepal The Nepali caste system resembles in some respects the Indian jāti system, with numerous jāti divisions with a varna system superimposed. Inscriptions attest the beginnings of a caste system during the Licchavi period. Jayasthiti Malla (1382–1395) categorised Newars into 64 castes (Gellner 2001). A similar exercise was made during the reign of Mahindra Malla (1506–1575). The Hindu social code was later set up in the Gorkha Kingdom by Ram Shah (1603–1636). Pakistan McKim Marriott claims a social stratification that is hierarchical, closed, endogamous and hereditary is widely prevalent, particularly in western parts of Pakistan. Frederik Barth in his review of this system of social stratification in Pakistan suggested that these are castes. Sri Lanka The caste system in Sri Lanka is a division of society into strata, influenced by the textbook varnas and jāti system found in India. Ancient Sri Lankan texts such as the Pujavaliya, Sadharmaratnavaliya and Yogaratnakaraya and inscriptional evidence show that the above hierarchy prevailed throughout the feudal period. The repetition of the same caste hierarchy even as recently as the 18th century, in the Kandyan-period Kadayimpoth – Boundary books as well indicates the continuation of the tradition right up to the end of Sri Lanka's monarchy. Outside South Asia Southeast Asia Indonesia Balinese caste structure has been described as being based either on three categories—the noble triwangsa (thrice born), the middle class of dwijāti (twice born), and the lower class of ekajāti (once born)--or on four castes Brahminas – priest Satrias – knighthood Wesias – commerce Sudras – servitude The Brahmana caste was further subdivided by Dutch ethnographers into two: Siwa and Buda. The Siwa caste was subdivided into five: Kemenuh, Keniten, Mas, Manuba and Petapan. This classification was to accommodate the observed marriage between higher-caste Brahmana men with lower-caste women. The other castes were similarly further sub-classified by 19th-century and early-20th-century ethnographers based on numerous criteria ranging from profession, endogamy or exogamy or polygamy, and a host of other factors in a manner similar to castas in Spanish colonies such as Mexico, and caste system studies in British colonies such as India. Philippines In the Philippines, pre-colonial societies do not have a single social structure. The class structures can be roughly categorised into four types: Classless societies – egalitarian societies with no class structure. Examples include the Mangyan and the Kalanguya peoples. Warrior societies – societies where a distinct warrior class exists, and whose membership depends on martial prowess. Examples include the Mandaya, Bagobo, Tagakaulo, and B'laan peoples who had warriors called the bagani or magani. Similarly, in the Cordillera highlands of Luzon, the Isneg and Kalinga peoples refer to their warriors as mengal or maingal. This society is typical for head-hunting ethnic groups or ethnic groups which had seasonal raids (mangayaw) into enemy territory. Petty plutocracies – societies which have a wealthy class based on property and the hosting of periodic prestige feasts. In some groups, it was an actual caste whose members had specialised leadership roles, married only within the same caste, and wore specialised clothing. These include the kadangyan of the Ifugao, Bontoc, and Kankanaey peoples, as well as the baknang of the Ibaloi people. In others, though wealth may give one prestige and leadership qualifications, it was not a caste per se. Principalities – societies with an actual ruling class and caste systems determined by birthright. Most of these societies are either Indianized or Islamized to a degree. They include the larger coastal ethnic groups like the Tagalog, Kapampangan, Visayan, and Moro societies. Most of them were usually divided into four to five caste systems with different names under different ethnic groups that roughly correspond to each other. The system was more or less feudalistic, with the datu ultimately having control of all the lands of the community. The land is subdivided among the enfranchised classes, the sakop or sa-op (vassals, lit. "those under the power of another"). The castes were hereditary, though they were not rigid. They were more accurately a reflection of the interpersonal political relationships, a person is always the follower of another. People can move up the caste system by marriage, by wealth, or by doing something extraordinary; and conversely they can be demoted, usually as criminal punishment or as a result of debt. Shamans are the exception, as they are either volunteers, chosen by the ranking shamans, or born into the role by innate propensity for it. They are enumerated below from the highest rank to the lowest: Royalty – (Visayan: kadatoan) the datu and immediate descendants. They are often further categorised according to purity of lineage. The power of the datu is dependent on the willingness of their followers to render him respect and obedience. Most roles of the datu were judicial and military. In case of an unfit datu, support may be withdrawn by his followers. Datu were almost always male, though in some ethnic groups like the Banwaon people, the female shaman (babaiyon) co-rules as the female counterpart of the datu. Nobility – (Visayan: tumao; Tagalog: maginoo; Kapampangan ginu; Tausug: bangsa mataas) the ruling class, either inclusive of or exclusive of the royal family. Most are descendants of the royal line or gained their status through wealth or bravery in battle. They owned lands and subjects, from whom they collected taxes. Shamans – (Visayan: babaylan; Tagalog: katalonan) the spirit mediums, usually female or feminised men. While they were not technically a caste, they commanded the same respect and status as nobility. Warriors – (Visayan: timawa; Tagalog: maharlika) the martial class. They could own land and subjects like the higher ranks, but were required to fight for the datu in times of war. In some Filipino ethnic groups, they were often tattooed extensively to record feats in battle and as protection against harm. They were sometimes further subdivided into different classes, depending on their relationship with the datu. They traditionally went on seasonal raids on enemy settlements. Commoners and slaves – (Visayan, Maguindanao: ulipon; Tagalog: alipin; Tausug: kiapangdilihan; Maranao: kakatamokan) – the lowest class composed of the rest of the community who were not part of the enfranchised classes. They were further subdivided into the commoner class who had their own houses, the servants who lived in the houses of others, and the slaves who were usually captives from raids, criminals, or debtors. Most members of this class were equivalent to the European serf class, who paid taxes and can be conscripted to communal tasks, but were more or less free to do as they please. East Asia China and Mongolia During the period of Yuan Dynasty, ruler Kublai Khan enforced a Four Class System, which was a legal caste system. The order of four classes of people in descending order were: Mongolian Semu people Han people (in the northern areas of China) Southerners (people of the former Southern Song dynasty) Today, the Hukou system is argued by various Western sources to be the current caste system of China. Tibet There is significant controversy over the social classes of Tibet, especially with regards to the serfdom in Tibet controversy. has put forth the argument that pre-1950s Tibetan society was functionally a caste system, in contrast to previous scholars who defined the Tibetan social class system as similar to European feudal serfdom, as well as non-scholarly western accounts which seek to romanticise a supposedly 'egalitarian' ancient Tibetan society. Japan In Japan's history, social strata based on inherited position rather than personal merit, were rigid and highly formalised in a system called mibunsei (身分制). At the top were the Emperor and Court nobles (kuge), together with the Shōgun and daimyō. Below them, the population was divided into four classes: samurai, peasants, craftsmen and merchants. Only samurai were allowed to bear arms. A samurai had a right to kill any peasants, craftsman or merchant who he felt were disrespectful. Merchants were the lowest caste because they did not produce any products. The castes were further sub-divided; for example, peasants were labelled as furiuri, tanagari, mizunomi-byakusho among others. As in Europe, the castes and sub-classes were of the same race, religion and culture. Howell, in his review of Japanese society notes that if a Western power had colonised Japan in the 19th century, they would have discovered and imposed a rigid four-caste hierarchy in Japan. De Vos and Wagatsuma observe that Japanese society had a systematic and extensive caste system. They discuss how alleged caste impurity and alleged racial inferiority, concepts often assumed to be different, are superficial terms, and are due to identical inner psychological processes, which expressed themselves in Japan and elsewhere. Endogamy was common because marriage across caste lines was socially unacceptable. Japan had its own untouchable caste, shunned and ostracised, historically referred to by the insulting term eta, now called burakumin. While modern law has officially abolished the class hierarchy, there are reports of discrimination against the buraku or burakumin underclasses. The burakumin are regarded as "ostracised". The burakumin are one of the main minority groups in Japan, along with the Ainu of Hokkaidō and those of Korean or Chinese descent. Korea The baekjeong (백정) were an "untouchable" outcaste of Korea. The meaning today is that of butcher. It originates in the Khitan invasion of Korea in the 11th century. The defeated Khitans who surrendered were settled in isolated communities throughout Goryeo to forestall rebellion. They were valued for their skills in hunting, herding, butchering, and making of leather, common skill sets among nomads. Over time, their ethnic origin was forgotten, and they formed the bottom layer of Korean society. In 1392, with the foundation of the Confucian Joseon dynasty, Korea systemised its own native class system. At the top were the two official classes, the Yangban, which literally means "two classes". It was composed of scholars (munban) and warriors (muban). Scholars had a significant social advantage over the warriors. Below were the jung-in (중인-中人: literally "middle people". This was a small class of specialised professions such as medicine, accounting, translators, regional bureaucrats, etc. Below that were the sangmin (상민-常民: literally 'commoner'), farmers working their own fields. Korea also had a serf population known as the nobi. The nobi population could fluctuate up to about one third of the population, but on average the nobi made up about 10% of the total population. In 1801, the vast majority of government nobi were emancipated, and by 1858 the nobi population stood at about 1.5% of the total population of Korea. The hereditary nobi system was officially abolished around 1886–87 and the rest of the nobi system was abolished with the Gabo Reform of 1894, but traces remained until 1930. The opening of Korea to foreign Christian missionary activity in the late 19th century saw some improvement in the status of the baekjeong. However, everyone was not equal under the Christian congregation, and even so protests erupted when missionaries tried to integrate baekjeong into worship, with non-baekjeong finding this attempt insensitive to traditional notions of hierarchical advantage. Around the same time, the baekjeong began to resist open social discrimination. They focused on social and economic injustices affecting them, hoping to create an egalitarian Korean society. Their efforts included attacking social discrimination by upper class, authorities, and "commoners", and the use of degrading language against children in public schools. With the Gabo reform of 1896, the class system of Korea was officially abolished. Following the collapse of the Gabo government, the new cabinet, which became the Gwangmu government after the establishment of the Korean Empire, introduced systematic measures for abolishing the traditional class system. One measure was the new household registration system, reflecting the goals of formal social equality, which was implemented by the loyalists' cabinet. Whereas the old registration system signified household members according to their hierarchical social status, the new system called for an occupation. While most Koreans by then had surnames and even bongwan, although still substantial number of cheonmin, mostly consisted of serfs and slaves, and untouchables did not. According to the new system, they were then required to fill in the blanks for surname in order to be registered as constituting separate households. Instead of creating their own family name, some cheonmins appropriated their masters' surname, while others simply took the most common surname and its bongwan in the local area. Along with this example, activists within and outside the Korean government had based their visions of a new relationship between the government and people through the concept of citizenship, employing the term inmin ("people") and later, kungmin ("citizen"). North Korea The Committee for Human Rights in North Korea reported that "Every North Korean citizen is assigned a heredity-based class and socio-political rank over which the individual exercises no control but which determines all aspects of his or her life." Called Songbun, Barbara Demick describes this "class structure" as an updating of the hereditary "caste system", a combination of Confucianism and Stalinism. It originated in 1946 and was entrenched by the 1960s, and consisted of 53 categories ranging across three classes: loyal, wavering, and impure. The privileged "loyal" class included members of the Korean Workers' Party and Korean People's Army officers' corps, the wavering class included peasants, and the impure class included collaborators with Imperial Japan and landowners. She claims that a bad family background is called "tainted blood", and that by law this "tainted blood" lasts three generations. West Asia Kurdistan Yazidis There are three hereditary groups, often called castes, in Yazidism. Membership in the Yazidi society and a caste is conferred by birth. Pîrs and Sheikhs are the priestly castes, which are represented by many sacred lineages (). Sheikhs are in charge of both religious and administrative functions and are divided into three endogamous houses, Şemsanî, Adanî and Qatanî who are in turn divided into lineages. The Pîrs are in charge of purely religious functions and traditionally consist of 40 lineages or clans, but approximately 90 appellations of Pîr lineages have been found, which may have been a result of new sub-lineages arising and number of clans increasing over time due to division as Yazidis settled in different places and countries. Division could occur in one family, if there were a few brothers in one clan, each of them could become the founder of their own Pîr sub-clan (). Mirîds are the lay caste and are divided into tribes, who are each affiliated to a Pîr and a Sheikh priestly lineage assigned to the tribe. Iran Pre-Islamic Sassanid society was immensely complex, with separate systems of social organisation governing numerous different groups within the empire. Historians believe society comprised four social classes, which linguistic analysis indicates may have been referred to collectively as "pistras". The classes, from highest to lowest status, were priests (), warriors (), secretaries (), and commoners (). Yemen In Yemen there exists a hereditary caste, the African-descended Al-Akhdam who are kept as perennial manual workers. Estimates put their number at over 3.5 million residents who are discriminated, out of a total Yemeni population of around 22 million. Africa Various sociologists have reported caste systems in Africa. The specifics of the caste systems have varied in ethnically and culturally diverse Africa, however the following features are common – it has been a closed system of social stratification, the social status is inherited, the castes are hierarchical, certain castes are shunned while others are merely endogamous and exclusionary. In some cases, concepts of purity and impurity by birth have been prevalent in Africa. In other cases, such as the Nupe of Nigeria, the Beni Amer of East Africa, and the Tira of Sudan, the exclusionary principle has been driven by evolving social factors. West Africa Among the Igbo of Nigeria – especially Enugu, Anambra, Imo, Abia, Ebonyi, Edo and Delta states of the country – scholar Elijah Obinna finds that the Osu caste system has been and continues to be a major social issue. The Osu caste is determined by one's birth into a particular family irrespective of the religion practised by the individual. Once born into Osu caste, this Nigerian person is an outcast, shunned and ostracised, with limited opportunities or acceptance, regardless of his or her ability or merit. Obinna discusses how this caste system-related identity and power is deployed within government, Church and indigenous communities. The osu class systems of eastern Nigeria and southern Cameroon are derived from indigenous religious beliefs and discriminate against the "Osus" people as "owned by deities" and outcasts. The Songhai economy was based on a caste system. The most common were metalworkers, fishermen, and carpenters. Lower caste participants consisted of mostly non-farm working immigrants, who at times were provided special privileges and held high positions in society. At the top were noblemen and direct descendants of the original Songhai people, followed by freemen and traders. In a review of social stratification systems in Africa, Richter reports that the term caste has been used by French and American scholars to many groups of West African artisans. These groups have been described as inferior, deprived of all political power, have a specific occupation, are hereditary and sometimes despised by others. Richter illustrates caste system in Ivory Coast, with six sub-caste categories. Unlike other parts of the world, mobility is sometimes possible within sub-castes, but not across caste lines. Farmers and artisans have been, claims Richter, distinct castes. Certain sub-castes are shunned more than others. For example, exogamy is rare for women born into families of woodcarvers. Similarly, the Mandé societies in Gambia, Ghana, Guinea, Ivory Coast, Liberia, Senegal and Sierra Leone have social stratification systems that divide society by ethnic ties. The Mande class system regards the jonow slaves as inferior. Similarly, the Wolof in Senegal is divided into three main groups, the geer (freeborn/nobles), jaam (slaves and slave descendants) and the underclass neeno. In various parts of West Africa, Fulani societies also have class divisions. Other castes include Griots, Forgerons, and Cordonniers. Tamari has described endogamous castes of over fifteen West African peoples, including the Tukulor, Songhay, Dogon, Senufo, Minianka, Moors, Manding, Soninke, Wolof, Serer, Fulani, and Tuareg. Castes appeared among the Malinke people no later than 14th century, and was present among the Wolof and Soninke, as well as some Songhay and Fulani populations, no later than 16th century. Tamari claims that wars, such as the Sosso-Malinke war described in the Sunjata epic, led to the formation of blacksmith and bard castes among the people that ultimately became the Mali empire. As West Africa evolved over time, sub-castes emerged that acquired secondary specialisations or changed occupations. Endogamy was prevalent within a caste or among a limited number of castes, yet castes did not form demographic isolates according to Tamari. Social status according to caste was inherited by off-springs automatically; but this inheritance was paternal. That is, children of higher caste men and lower caste or slave concubines would have the caste status of the father. Central Africa Ethel M. Albert in 1960 claimed that the societies in Central Africa were caste-like social stratification systems. Similarly, in 1961, Maquet notes that the society in Rwanda and Burundi can be best described as castes. The Tutsi, noted Maquet, considered themselves as superior, with the more numerous Hutu and the least numerous Twa regarded, by birth, as respectively, second and third in the hierarchy of Rwandese society. These groups were largely endogamous, exclusionary and with limited mobility. Horn of Africa In a review published in 1977, Todd reports that numerous scholars report a system of social stratification in different parts of Africa that resembles some or all aspects of caste system. Examples of such caste systems, he claims, are to be found in Ethiopia in communities such as the Gurage and Konso. He then presents the Dime of Southwestern Ethiopia, amongst whom there operates a system which Todd claims can be unequivocally labelled as caste system. The Dime have seven castes whose size varies considerably. Each broad caste level is a hierarchical order that is based on notions of purity, non-purity and impurity. It uses the concepts of defilement to limit contacts between caste categories and to preserve the purity of the upper castes. These caste categories have been exclusionary, endogamous and the social identity inherited. Alula Pankhurst has published a study of caste groups in SW Ethiopia. Among the Kafa, there were also traditionally groups labelled as castes. "Based on research done before the Derg regime, these studies generally presume the existence of a social hierarchy similar to the caste system. At the top of this hierarchy were the Kafa, followed by occupational groups including blacksmiths (Qemmo), weavers (Shammano), bards (Shatto), potters, and tanners (Manno). In this hierarchy, the Manjo were commonly referred to as hunters, given the lowest status equal only to slaves." The Borana Oromo of southern Ethiopia in the Horn of Africa also have a class system, wherein the Wata, an acculturated hunter-gatherer group, represent the lowest class. Though the Wata today speak the Oromo language, they have traditions of having previously spoken another language before adopting Oromo. The traditionally nomadic Somali people are divided into clans, wherein the Rahanweyn agro-pastoral clans and the occupational clans such as the Madhiban were traditionally sometimes treated as outcasts. As Gabboye, the Madhiban along with the Yibir and Tumaal (collectively referred to as sab) have since obtained political representation within Somalia, and their general social status has improved with the expansion of urban centers. Europe European feudalism with its rigid aristocracy can also be considered as a caste system. Basque region For centuries, through the modern times, the majority regarded Cagots who lived primarily in the Basque region of France and Spain as an inferior caste, the untouchables. While they had the same skin color and religion as the majority, in the churches they had to use segregated doors, drink from segregated fonts, and receive communion on the end of long wooden spoons. It was a closed social system. The socially isolated Cagots were endogamous, and chances of social mobility non-existent. United Kingdom In July 2013, the UK government announced its intention to amend the Equality Act 2010, to "introduce legislation on caste, including any necessary exceptions to the caste provisions, within the framework of domestic discrimination law". Section 9(5) of the Equality Act 2010 provides that "a Minister may by order amend the statutory definition of race to include caste and may provide for exceptions in the Act to apply or not to apply to caste". From September 2013 to February 2014, Meena Dhanda led a project on "Caste in Britain" for the UK Equality and Human Rights Commission (EHRC). Americas Latin America The existence of a caste system based on the concept of casta in Latin America under colonial Spain has been raised and contemporarily contested, distinguishing it from more general colonial or racial discrimination. United States A survey on caste discrimination conducted by Equality Labs found 67% of Indian Dalits living in the US reporting that they faced caste-based harassment at the workplace, and 27% reporting verbal or physical assault based on their caste. In 2023, Seattle became the first city in the United States to ban discrimination based on caste. In the opinion of W. Lloyd Warner, discrimination in the Southern United States in the 1930s against Blacks was similar to Indian castes in such features as residential segregation and marriage restrictions. In her 2020 book Caste: The Origins of Our Discontents, journalist Isabel Wilkerson used caste as an analogy to understand racial discrimination in the United States. Gerald D. Berreman contrasted the differences between discrimination in the United States and India. In India, there are complex religious features which make up the system, whereas in the United States race and color are the basis for differentiation. The caste systems in India and the United States have higher groups which desire to retain their positions for themselves and thus perpetuate the two systems. The process of creating a homogenized society by social engineering in both India and the Southern US has created other institutions that have made class distinctions among different groups evident. Anthropologist James C. Scott elaborates on how "global capitalism is perhaps the most powerful force for homogenization, whereas the state may be the defender of local difference and variety in some instances". The caste system, a relic of feudalistic economic systems, emphasizes differences between socio-economic classes that are obviated by openly free market capitalistic economic systems, which reward individual initiative, enterprise, merit, and thrift, thereby creating a path for social mobility. When the feudalistic slave economy of the southern United States was dismantled, Jim Crow laws and acts of domestic terrorism committed by white supremacists prevented many industrious African Americans from participating in the formal economy and achieving economic success on parity with their white peers, or destroying that economic success in instances where it was achieved, such as Black wall street, with only rare but commonly touted exceptions of lasting personal success such as Maggie Walker, Annie Malone, and Madame C.J. Walker. Parts of the United States are sometimes divided by race and class status despite the national narrative of integration. Caste in sociology and entomology The initial observational studies of the division of labour in ant colonies attempted to demonstrate that ants specialized in tasks that were best suited to their size when they emerged from the pupae stage into the adult stage. A large proportion of the experimental work was done in species that showed strong variation in size. As the size of an adult was fixed for life, workers of a specific size range came to be called a "caste", calling up the traditional caste system in India in which a human's standing in society was decided at birth. The notion of caste encouraged a link between scholarship in entomology and sociology because it served as an example of a division of labour in which the participants seemed to be uncompromisingly adapted to special functions and sometimes even unique environments. To bolster the concept of caste, entomologists and sociologists referred to the complementary social or natural parallel and thereby appeared to generalize the concept and give it an appearance of familiarity. In the late 19th- and early 20th centuries, the perceived similarities between the Indian caste system and caste polymorphism in insects were used to create a correspondence or parallelism for the purpose of explaining or clarifying racial stratification in human societies; the explanations came particularly to be employed in the United States. Ideas from heredity and natural selection influenced some sociologists who believed that some groups were predetermined to belong to a lower social or occupational status. Chiefly through the work of W. Lloyd Warner at the University of Chicago, a group of sociologists sharing similar principles came to evolve around the creed of caste in the 1930s and 1940s. The ecologically-oriented sociologist Robert E. Park, although attributing more weight to environmental explanations than the biological nonetheless believed that there were obstacles to the assimilation of blacks into American society and that an "accommodation stage" in a biracially organized caste system was required before full assimilation. He did disavow his position in 1937, suggesting that blacks were a minority and not a caste. The Indian sociologist Radhakamal Mukerjee was influenced by Robert E. Park and adopted the concept of "caste" to describe race relations in the US. According to anthropologist Diane Rodgers, Mukerjee "proceeded to suggest that a caste system should be correctly instituted in the (US) South to ease race relations." Mukerjee often employed both entomological and sociological data and clues to describe caste systems. He wrote "while the fundamental industries of man are dispersed throughout the insect world, the same kind of polymorphism appears again and again in different species of social insects which have reacted in the same manner as man, under the influence of the same environment, to ensure the supply and provision of subsistence." Comparing the caste system in India to caste polymorphism in insects, he noted, "where we find the organization of social insects developed to perfection, there also has been seen among human associations a minute and even rigid specialization of functions, along with ant- and bee-like societal integrity and cohesiveness." He considered the "resemblances between insect associations and caste-ridden societies" to be striking enough to be "amusing". See also Estates of the realm Inter-caste marriages in India Job Kamaiya Priestly caste Propiska Social exclusion Warrior caste Notes References Sources Oxford English Dictionary () Quote: caste, n. 2a. spec. One of the several hereditary classes into which society in India has from time immemorial been divided; ... This is now the leading sense, which influences all others. Further reading Spectres of Agrarian Territory by David Ludden 11 December 2001 "Early Evidence for Caste in South India", pp. 467–492 in Dimensions of Social Life: Essays in honour of David G. Mandelbaum, Edited by Paul Hockings and Mouton de Gruyter, Berlin, New York, Amsterdam, 1987. External links Casteless Auguste Comte on why and how castes developed across the world – in The Positive Philosophy, Volume 3 (see page 55 onwards) Robert Merton on Caste and The Sociology of Science Caste, Society and Politics in India from the Eighteenth Century to the Modern Age – Susan Bayly Class In Yemen by Marguerite Abadjian (Archive of the Baltimore Sun) International Dalit Solidarity Network: An international advocacy group for Dalits Social status
7451
https://en.wikipedia.org/wiki/Jews%20as%20the%20chosen%20people
Jews as the chosen people
In Judaism, the concept of the Jews as the chosen people ( ha-ʿam ha-nivḥar , IPA: haʕam hanivħar) is the belief that the Jews, via descent from the ancient Israelites, are the chosen people, i.e. selected to be in a covenant with God. The idea of the Israelites being chosen by God is found most directly in the Book of Deuteronomy as the verb baḥar (בָּחַר), and is alluded to elsewhere in the Hebrew Bible using other terms such as "holy people". Much is written about these topics in rabbinic literature. The three largest Jewish denominations—Orthodox Judaism, Conservative Judaism and Reform Judaism—maintain the belief that the Jews have been chosen by God for a purpose. Sometimes this choice is seen as charging the Jewish people with a specific mission—to be a light unto the nations, and to exemplify the covenant with God as described in the Torah. While the concept of "chosenness" may be understood by some to connote ethnic supremacy, the status as a "chosen people" within Judaism does not preclude a belief that God has a relationship with other peoples—rather, Judaism holds that God had entered into a covenant with all humankind, and that Jews and non-Jews alike have a relationship with God. Biblical references as well as rabbinic literature support this view: Moses refers to the "God of the spirits of all flesh", the Tanakh (Hebrew Bible) also identifies prophets outside the community of Israel and the prophet Yonah (Jonah) is explicitly told to go prophesize to the non-Jewish people of Nineveh. Based on these statements and stories, some rabbis theorized that, in the words of Natan'el al-Fayyumi, a Yemenite Jewish theologian of the 12th century, "God permitted to every people something he forbade to others ... [and] God sends a prophet to every people according to their own language." (Levine, 1907/1966) The Mishnah states that "Humanity was produced from one man, Adam, to show God's greatness. When a man mints a coin in a press, each coin is identical. But when the King of Kings, the Holy One, blessed be He, creates people in the form of Adam not one is similar to any other" (Mishnah Sanhedrin 4:5). The Tosefta, an important supplement to the Mishnah, also states: "Righteous people of all nations have a share in the world to come" (Sanhedrin 105a). According to the Israel Democracy Institute, approximately two thirds of Israeli Jews believe that Jews are the "chosen people". In the Bible According to Deuteronomy, when the Lord delivers the Israelites to the land, the other nations will be cast out, and "thou shalt make no covenant with them, nor show mercy unto them" Deuteronomy 7:5-7:6, "But thus shall ye deal with them; ye shall destroy their altars, and break down their images, and cut down their groves, and burn their graven images with fire. For thou art a holy people unto the Lord thy God: the Lord thy God hath chosen thee to be a special people unto himself, above all people that are upon the face of the earth". A similar passage exalting Israel as the chosen people follows prohibitions on baldness and cutting yourself in mourning, "For thou art a holy people". The Torah also says, "Now therefore, if you will obey my voice indeed, and keep my covenant, then you shall be a peculiar treasure unto me from all the peoples, for all the earth is mine." God promises that he will never exchange his people with any other: "And I will establish My covenant between Me and you and your descendants after you in their generations, for an everlasting covenant, to be God to you and your descendants after you." Other Torah verses about chosenness, "And you shall be unto me a kingdom of priests, and a holy nation" "The Lord did not set his love upon you, nor choose you, because you were more in number than any people; for you were the fewest of all people; but because the Lord loved you, and because he would keep the oath which he had sworn unto your ancestors." The obligation imposed upon the Israelites was emphasized by the prophet Amos: "You only have I singled out of all the families of the earth: therefore will I visit upon you all your iniquities." Rabbinic views Sometimes this choice is seen as charging the Jewish people with a specific mission—to be a light unto the nations, and to exemplify the covenant with God as described in the Torah. This view, however, does not always preclude a belief that God has a relationship with other peoples—rather, Judaism held that God had entered into a covenant with all humankind, and that Jews and non-Jews alike have a relationship with God. Biblical references as well as rabbinic literature support this view: Moses refers to the "God of the spirits of all flesh", and the Tanakh also identifies prophets outside the community of Israel. Based on these statements, some rabbis theorized that, in the words of Natan'el al-Fayyumi, a Yemenite Jewish theologian of the 12th century, "God permitted to every people something he forbade to others...[and] God sends a prophet to every people according to their own language." The Mishnah states that "Humanity was produced from one man, Adam, to show God's greatness. When a man mints a coin in a press, each coin is identical. But when the King of Kings, the Holy One, blessed be He, creates people in the form of Adam not one is similar to any other." The Tosefta, a collection of important post-Talmudic discourses, also states: "Righteous people of all nations have a share in the world to come." Most Jewish texts do not state that "God chose the Jews" by itself. Rather, this is usually linked with a mission or purpose, such as proclaiming God's message among all the nations, even though Jews cannot become "unchosen" if they shirk their mission. This implies a special duty, which evolves from the belief that Jews have been pledged by the covenant which God concluded with the biblical patriarch Abraham, their ancestor, and again with the entire Jewish nation at Mount Sinai. In this view, Jews are charged with living a holy life as God's priest-people. In the Jewish prayerbook (the Siddur), chosenness is referred to in a number of ways. The blessing for reading the Torah reads, "Praised are You, Lord our God, King of the Universe, Who has chosen us out of all the nations and bestowed upon us His Torah." In the "Kiddush", a prayer of sanctification, in which the Sabbath is inaugurated over a cup of wine, the text reads, "For you have chosen us and sanctified us out of all the nations, and have given us the Sabbath as an inheritance in love and favour. Praised are you, Lord, who hallows the Sabbath." In the "Kiddush" recited on festivals it reads, "Blessed are You ... who have chosen us from among all nations, raised us above all tongues, and made us holy through His commandments." The Aleinu prayer refers to the concept of Jews as a chosen people: "It is our duty to praise the Master of all, to exalt the Creator of the Universe, who has not made us like the nations of the world and has not placed us like the families of the earth; who has not designed our destiny to be like theirs, nor our lot like that of all their multitude. We bend the knee and bow and acknowledge before the Supreme King of Kings, the Holy One, blessed be he, that it is he who stretched forth the heavens and founded the earth. His seat of glory is in the heavens above; his abode of majesty is in the lofty heights. Further interpretations The following section contains information from the Jewish Encyclopedia, originally published between 1901–1906, which is in the public domain. According to the Rabbis, "Israel is of all nations the most willful or headstrong one, and the Torah was to give it the right scope and power of resistance, or else the world could not have withstood its fierceness." "The Lord offered the Law to all nations; but all refused to accept it except Israel." How do we understand "A Gentile who consecrates his life to the study and observance of the Law ranks as high as the high priest", says R. Meïr, by deduction from Lev. xviii. 5; II Sam. vii. 19; Isa. xxvi. 2; Ps. xxxiii. 1, cxviii. 20, cxxv. 4, where all stress is laid not on Israel, but on man or the righteous one. Maimonides states: It is now abundantly clear that the pledges Hashem made to Avraham and his descendants would be fulfilled exclusively first in Yitzchak and then in Yaakov, Yitzchak son. This is confirmed by a passage that states, "He is ever mindful of His covenant ... that He made with Avraham, swore to Yitzchak, and confirmed in a decree for Yaakov, for Yisrael, as an eternal covenant." The Gemara states this regarding a non-Jew who studies Torah [his 7 mitzvot] and regarding this, see Shita Mekubetzes, Bava Kama 38a who says that this is an exaggeration. In any case, this statement was not extolling the non-Jew. The Rishonim explain that it is extolling the Torah. Tosfos explains that it uses the example of a kohen gadol (high priest), because this statement is based on the verse, "y'kara hi mipnimim" (it is more precious than pearls). This is explained elsewhere in the Gemara to mean that the Torah is more precious pnimim (translated here as "inside" instead of as "pearls"; thus that the Torah is introspectively absorbed into the person), which refers to lifnai v'lifnim (translated as "the most inner of places"), that is the Holy of Holies where the kahon gadol went. In any case, in Midrash Rabba this statement is made with an important addition: a non-Jew who converts and studies Torah etc. The Nation of Israel is likened to the olive. Just as this fruit yields its precious oil only after being much pressed and squeezed, so Israel's destiny is one of great oppression and hardship, in order that it may thereby give forth its illuminating wisdom. Poverty is the quality most befitting Israel as the chosen people. Only on account of its good works is Israel among the nations "as the lily among thorns", or "as wheat among the chaff." Modern Orthodox views Rabbi Lord Immanuel Jakobovits, former Chief Rabbi of the United Synagogue of Great Britain (Modern Orthodox Judaism), described chosenness in this way: "Yes, I do believe that the chosen people concept as affirmed by Judaism in its holy writ, its prayers, and its millennial tradition. In fact, I believe that every people—and indeed, in a more limited way, every individual—is "chosen" or destined for some distinct purpose in advancing the designs of Providence. Only, some fulfill their mission and others do not. Maybe the Greeks were chosen for their unique contributions to art and philosophy, the Romans for their pioneering services in law and government, the British for bringing parliamentary rule into the world, and the Americans for piloting democracy in a pluralistic society. The Jews were chosen by God to be 'peculiar unto Me' as the pioneers of religion and morality; that was and is their national purpose." Modern Orthodox theologian Michael Wyschogrod wrote: "[T]he initial election of Abraham himself was not earned. ... We are simply told that God commanded Abraham to leave his place of birth and go to a land that God would show him. He is also promised that his descendants will become a numerous people. But nowhere does the Bible tell us why Abraham rather than someone else was chosen. The implication is that God chooses whom He wishes and that He owes no accounting to anyone for His choices." Rabbi Norman Lamm, a leader of Modern Orthodox Judaism, wrote: "The chosenness of Israel relates exclusively to its spiritual vocation embodied in the Torah; the doctrine, indeed, was announced at Sinai. Whenever it is mentioned in our liturgy—such as the blessing immediately preceding the Shema....it is always related to Torah or Mitzvot (commandments). This spiritual vocation consists of two complementary functions, described as "Goy Kadosh", that of a holy nation, and "Mamlekhet Kohanim", that of a kingdom of priests. The first term denotes the development of communal separateness or differences in order to achieve a collective self-transcendence. ... The second term implies the obligation of this brotherhood of the spiritual elite toward the rest of mankind; priesthood is defined by the prophets as fundamentally a teaching vocation." Conservative views Conservative Judaism views the concept of chosenness in this way: "Few beliefs have been subject to as much misunderstanding as the 'Chosen People' doctrine. The Torah and the Prophets clearly stated that this does not imply any innate Jewish superiority. In the words of Amos (3:2) 'You alone have I singled out of all the families of the earth—that is why I will call you to account for your iniquities.' The Torah tells us that we are to be "a kingdom of priests and a holy nation" with obligations and duties which flowed from our willingness to accept this status. Far from being a license for special privilege, it entailed additional responsibilities not only toward God but to our fellow human beings. As expressed in the blessing at the reading of the Torah, our people have always felt it to be a privilege to be selected for such a purpose. For the modern traditional Jew, the doctrine of the election and the covenant of Israel offers a purpose for Jewish existence which transcends its own self interests. It suggests that because of our special history and unique heritage we are in a position to demonstrate that a people that takes seriously the idea of being covenanted with God can not only thrive in the face of oppression, but can be a source of blessing to its children and its neighbors. It obligates us to build a just and compassionate society throughout the world and especially in the land of Israel where we may teach by example what it means to be a 'covenant people, a light unto the nations.'" Rabbi Reuven Hammer comments on the excised sentence in the Aleinu prayer mentioned above: "Originally the text read that God has not made us like the nations who "bow down to nothingness and vanity, and pray to an impotent god", ... In the Middle Ages these words were censored, since the church believed they were an insult to Christianity. Omitting them tends to give the impression that the Aleinu teaches that we are both different and better than others. The actual intent is to say that we are thankful that God has enlightened us so that, unlike the pagans, we worship the true God and not idols. There is no inherent superiority in being Jewish, but we do assert the superiority of monotheistic belief over paganism. Although paganism still exists today, we are no longer the only ones to have a belief in one God." Reform views Reform Judaism views the concept of chosenness as follows: "Throughout the ages it has been Israel's mission to witness to the Divine in the face of every form of paganism and materialism. We regard it as our historic task to cooperate with all men in the establishment of the kingdom of God, of universal brotherhood, Justice, truth and peace on earth. This is our Messianic goal." In 1999 the Reform movement stated, "We affirm that the Jewish people are bound to God by an eternal covenant, as reflected in our varied understandings of Creation, Revelation and Redemption. ... We are Israel, a people aspiring to holiness, singled out through our ancient covenant and our unique history among the nations to be witnesses to God's presence. We are linked by that covenant and that history to all Jews in every age and place." Alternative views Equality of souls Many Kabbalistic sources, notably the Tanya, contain statements to the effect that the Jewish soul is qualitatively different from the non-Jewish soul. A number of known Chabad rabbis offered alternative readings of the Tanya, did not take this teaching literally, and even managed to reconcile it with the leftist ideas of internationalism and class struggle. The original text of the Tanya refers to the "idol worshippers" and does not mention the "nations of the world" at all, although such interpretation was endorsed by Menachem Mendel Schneerson and is popular in contemporary Chabad circles. Hillel of Parich, an early Tanya commentator, wrote that the souls of righteous Gentiles are more similar to the Jewish souls, and are generally good and not egoistic. This teaching was accepted by Schneerson and is considered normative in Chabad. Different in character but not different in value According to the author of the Tanya himself, a righteous non-Jew can achieve a high level of spiritually, similar to an angel, although his soul is still fundamentally different in character, but not value, from a Jewish one. Tzemach Tzedek, the third rebbe of Chabad, wrote that the Muslims are naturally good-hearted people. Rabbi Yosef Jacobson, a popular contemporary Chabad lecturer, teaches that in today's world most non-Jews belong to the category of righteous Gentiles, effectively rendering the Tanya's attitude anachronistic. Altruism An anti-Zionist interpretation of Tanya was offered by Abraham Yehudah Khein, a prominent Ukrainian Chabad rabbi, who supported anarchist communism and considered Peter Kropotkin a great Tzaddik. Khein basically read the Tanya backwards; since the souls of idol worshipers are known to be evil, according to the Tanya, while the Jewish souls are known to be good, he concluded that truly altruistic people are really Jewish, in a spiritual sense, while Jewish nationalists and class oppressors are not. By this logic, he claimed that Vladimir Solovyov and Rabindranath Tagore probably have Jewish souls, while Leon Trotsky and other totalitarians do not, and many Zionists, whom he compared to apes, are merely "Jewish by birth certificate". Righteous non-Jews Nachman of Breslov also believed that Jewishness is a level of consciousness, and not an intrinsic inborn quality. He wrote that, according to the Book of Malachi, one can find "potential Jews" among all nations, whose souls are illuminated by the leap of "holy faith", which "activated" the Jewishness in their souls. These people would otherwise convert to Judaism, but prefer not to do so. Instead, they recognize the Divine unity within their pagan religions. Isaac Arama, an influential philosopher and mystic of the 15th century, believed that righteous non-Jews are spiritually identical to the righteous Jews. Rabbi Menachem Meiri, a famous Catalan Talmudic commentator and Maimonidian philosopher, considered all people, who sincerely profess an ethical religion, to be part of a greater "spiritual Israel". He explicitly included Christians and Muslims in this category. Meiri rejected all Talmudic laws that discriminate between the Jews and non-Jews, claiming that they only apply to the ancient idolators, who had no sense of morality. The only exceptions are a few laws related directly or indirectly to intermarriage, which Meiri did recognize. Meiri applied his idea of "spiritual Israel" to the Talmudic statements about unique qualities of the Jewish people. For example, he believed that the famous saying that Israel is above astrological predestination (Ein Mazal le-Israel) also applied to the followers of other ethical faiths. He also considered countries, inhabited by decent moral non-Jews, such as Languedoc, as a spiritual part of the Holy Land. Spinoza One Jewish critic of chosenness was the philosopher Baruch Spinoza. In the third chapter of his Theologico-Political Treatise, Spinoza mounts an argument against a naive interpretation of God's choice of the Jews. Bringing evidence from the Bible itself, he argues that God's choice of Israel was not unique (he had chosen other nations before choosing the Hebrew nation) and that the choice of the Jews is neither inclusive (it does not include all of the Jews, but only the 'pious' ones) nor exclusive (it also includes 'true gentile prophets'). Finally, he argues that God's choice is not unconditional. Recalling the numerous times God threatened the complete destruction of the Hebrew nation, he asserts that this choice is neither absolute, nor eternal, nor necessary. Einstein In a German-language letter to philosopher Eric Gutkind, dated 3 January 1954, the physicist Albert Einstein wrote: The word God is for me nothing more than the expression and product of human weaknesses, the Bible a collection of honorable, but still primitive legends which are nevertheless pretty childish. No interpretation no matter how subtle can (for me) change this.... For me the Jewish religion like all other religions is an incarnation of the most childish superstitions. And the Jewish people to whom I gladly belong and with whose mentality I have a deep affinity have no different quality for me than all other people.... I cannot see anything “chosen” about them. Reconstructionist criticism Reconstructionist Judaism rejects the concept of chosenness. Its founder, Rabbi Mordecai Kaplan, said that the idea that God chose the Jewish people leads to racist beliefs among Jews, and thus must be excised from Jewish theology. This rejection of chosenness is made explicit in the movement's siddurim (prayer books). For example, the original blessing recited before reading from the Torah contains the phrase, "asher bahar banu mikol ha’amim"—"Praised are you Lord our God, ruler of the Universe, who has chosen us from among all peoples by giving us the Torah." The Reconstructionist version is rewritten as "asher kervanu la’avodato", "Praised are you Lord our God, ruler of the Universe, who has drawn us to your service by giving us the Torah." In the mid-1980s, the Reconstructionist movement issued its Platform on Reconstructionism. It states that the idea of chosenness is "morally untenable", because anyone who has such beliefs "implies the superiority of the elect community and the rejection of others." Not all Reconstructionists accept this view. The newest siddur of the movement, Kol Haneshamah, includes the traditional blessings as an option, and some modern Reconstructionist writers have opined that the traditional formulation is not racist, and should be embraced. An original prayer book by Reconstructionist feminist poet Marcia Falk, The Book of Blessings, has been widely accepted by both Reform and Reconstructionist Jews. Falk rejects all concepts relating to hierarchy or distinction; she sees any distinction as leading to the acceptance of other kinds of distinctions, thus leading to prejudice. She writes that as a politically liberal feminist, she must reject distinctions made between men and women, homosexuals and heterosexuals, Jews and non-Jews, and to some extent even distinctions between the Sabbath and the other six days of the week. She thus rejects the idea of chosenness as unethical. She also rejects Jewish theology in general, and instead holds to a form of religious humanism. Falk writes: "The idea of Israel as God's chosen people ... is a key concept in rabbinic Judaism. Yet it is particularly problematic for many Jews today, in that it seems to fly in the face of monotheistic belief that all humanity is created in the divine image—and hence, all humanity is equally loved and valued by God. ... I find it difficult to conceive of a feminist Judaism that would incorporate it in its teaching: the valuing of one people over and above others is all too analogous to the privileging of one sex over another." Reconstructionist author Judith Plaskow also criticises the idea of chosenness, for many of the same reasons as Falk. A politically liberal lesbian, Plaskow rejects most distinctions made between men and women, homosexuals and heterosexuals, and Jews and non-Jews. In contrast to Falk, Plaskow does not reject all concepts of difference as inherently leading to unethical beliefs, and holds to a more classical form of Jewish theism than Falk. A number of responses to these views have been made by Reform and Conservative Jews; they hold that these criticisms are against teachings that do not exist within liberal forms of Judaism, and which are rare in Orthodox Judaism (outside certain Haredi communities, such as Chabad). A separate criticism stems from the very existence of feminist forms of Judaism in all denominations of Judaism, which do not have a problem with the concepts of chosenness. Views of other religions Islam The children of Israel enjoy a special status in the Islamic book, the Quran (2:47 and 2:122). However, Muslim scholars point out that this status did not confer upon Israelites any racial superiority, and was only valid so long as the Israelites maintain their covenant with God. Christianity Some Christians believe that the Jews were God's chosen people, but because of Jewish Rejection of Jesus, the Christians in turn received that special status. This doctrine is known as Supersessionism. Other Christians, such as the Christadelphians, believe that God has not rejected Israel as his chosen people and that the Jews will in fact accept Jesus as their Messiah at his Second Coming, resulting in their salvation. Augustine criticized Jewish chosenness as "carnal." He reasoned that Israel was chosen "according to the flesh." Influence on Judaism's relationship with other religions Avi Beker, an Israeli scholar and former Secretary General of the World Jewish Congress, regarded the idea of the chosen people as Judaism's defining concept and "the central unspoken psychological, historical, and theological problem which is at the heart of Jewish-Gentile relations." In his book The Chosen: The History of an Idea, and the Anatomy of an Obsession, Beker expresses the view that the concept of chosenness is the driving force behind Jewish-Gentile relations, explaining both the admiration and, more pointedly, the envy and the hatred which the world has felt towards the Jews in both religious and secular terms. Beker argues that while Christianity has modified its doctrine on the displacement of the Jews, Islam has neither reversed nor reformed its theology concerning the succession of both the Jews and the Christians. According to Beker, this presents a major barrier to conflict resolution in the Arab-Israeli conflict. Ethnocentrism The Israeli philosopher Ze'ev Levy writes that chosenness can be "(partially) justified only from the historical angle" with respect to its spiritual and moral contribution to Jewish life through the centuries as "a powerful agent of consolation and hope". He points out, however, that modern anthropological theories "do not merely proclaim the inherent universal equality of all people [as] human beings; they also stress the equivalence [emphasis in original] of all human cultures." He continues that "there are no inferior and superior people or cultures but only different, other, ones." He concludes that the concept of chosenness entails ethnocentrism, "which does not go hand in hand with otherness, that is, with unconditional respect of otherness". Some people have claimed that Judaism's chosen people concept is racist because it implies that Jews are superior to non-Jews. The Anti-Defamation League asserts that the concept of a chosen people within Judaism has nothing to do with racial superiority. See also Religious nationalism Supersessionism Notes References Emet Ve-Emunah: Statement of Principles of Conservative Judaism, JTSA, New York, 1988, p. 33–34 Platform on Reconstructionism Federation of Reconstructionist Congregations and Havurot, September 1986, pages D, E Statement of Principles for Reform Judaism, 1999 Pittsburgh convention of the Central Conference of American Rabbis Encyclopaedia Judaica, Keter Publishing Ismar Elbogen Jewish Liturgy: A Comprehensive History JPS, 1993. The most thorough academic study of the Jewish liturgy ever written. Marcia Falk The Book of Blessings Harper San Francisco, 1996 Reuven Hammer, Ed. Or Hadash: A Commentary on Siddur Sim Shalom for Shabbat and Festivals, The Rabbinical Assembly, 2003 Nosson Scherman, Ed. The Complete Artscroll Siddur, Mesorah Publications, 2nd edition, 1986 Further reading (Part 1. Chosenness) External links The Contemporary Rivalry over the Chosen People: Jewish, Christian, and Muslim Perspectives – Avi Beker at the Institute for Global Jewish Affairs Chosen people at the Jewish Encyclopedia Beliefs of Reform Judaism The Jewish concept of chosenness When God Wanted to Destroy the Chosen People, Dr Gili Kugler The Chosen People FAQs Some are Chosen, All are Loved. Rabbi Gilbert S. Rosenthal The Chosen People: Reclaiming Our Sacred Myth, Mitchell Max How does Jewish Pride differ from Nazi Supremacy?, Rabbi Tzvi Freeman Anti-Defamation League paper on Christian Identity The Covenant, the Sword, and the Arm of the Lord – a Christian Identity movement The Real Truth About The Talmud by Rabbi Gil Student. Exposes fraudulent or distorted Talmud quotes used by antisemites Are the Jews the Chosen People? chabad.org OzTorah – Isn't it arrogant and exclusionist for Jews to regard themselves as the Chosen People? Jewish theology
7602
https://en.wikipedia.org/wiki/The%20Family%20International
The Family International
The Family International (TFI) is an American new religious movement founded in 1968 by David Brandt Berg. The group has gone under a number of different names since its inception, including Teens for Christ, The Children of God (COG), The Family of Love, or simply The Family. Former members have accused TFI of being an authoritarian cult which engages in the systematic physical and sexual abuse of children, resulting in lasting trauma among survivors. The group has also been accused of targeting vulnerable people. Overview According to the Canadian Broadcasting Corporation, "at its height" the Family movement had "tens of thousands of members, including River and Joaquin Phoenix, Rose McGowan and Jeremy Spencer". TFI initially spread a message of salvation, apocalypticism, spiritual "revolution and happiness" and distrust of the outside world, which the members called The System. Like some other fundamentalist groups, it "foretold the coming of a dictator called the anti-Christ, the rise of a brutal One World Government and its eventual overthrow by Jesus Christ, in the Second Coming". In 1976, it began a method of evangelism called Flirty Fishing that used sex to "show God's love and mercy" and win converts, resulting in controversy. TFI's founder and prophetic leader, David Berg (who was first called "Moses David" in the Texas press, and was also referred to "Father David" by members), gave himself the titles of "King", "The Last Endtime Prophet", "Moses", and "David". Berg communicated with his followers via "Mo Letters"—letters of instruction and counsel on myriad spiritual and practical subjects—until his death in late 1994. After his death, his widow Karen Zerby became the leader of TFI, taking the titles of "Queen" and "Prophetess". Zerby married Steve Kelly (also known as Peter Amsterdam), an assistant of Berg's whom Berg had handpicked as her "consort". Kelly took the title of "King Peter" and became the face of TFI, speaking in public more often than either Berg or Zerby. There have been multiple allegations of child sexual abuse made by past members. Berg preached a combination of traditional Christian evangelism, with elements popular with the Counterculture of the 1960s. There was much "end-of-the-world imagery" found in the Book of Revelation of the New Testament, preaching of impending doom for America and the ineffectiveness of established churches. Berg "urged a return to the early Christian community described in the Bible's Book of Acts, in which believers lived together and shared all", resembling communal living of late 1960s hippies. History The Children of God (1968–1977) The founder of the movement, David Brandt Berg (1919–1994), was a former Christian and Missionary Alliance pastor. Berg started in 1968 as an evangelical preacher with a following of "born-again hippies" who gathered at a coffeehouse in Huntington Beach, in Orange County, California. In 1969, after having a revelation "that California would be hit by a major earthquake", he left Huntington Beach and "took his followers on the road". They would proselytize in the streets and distribute pamphlets. Leaders within COG were referred to as The Chain. Members of The Children of God (COG) founded communes, first called colonies (now referred to as homes), in various cities. Berg communicated with his followers by writing letters. He published nearly 3,000 letters over a period of 24 years, referred to as the Mo Letters. In a letter written in January 1972, Berg stated that he was God's prophet for the contemporary world, attempting to further solidify his spiritual authority within the group. Berg's letters also contained public acknowledgement of his own failings and weaknesses, (for example, he issued a Mo Letter entitled "My confession -- I was an alcoholic!" (ML #1406 Summer 1982) relating his depression after some of his closest supporters quit in 1978. In 1972, a Mo Letter reportedly entitled "Flee as a Bird to Your Mountain" was interpreted by some members (such as Ruth Gordon) as a warning to leave America. "God was going to destroy the U.S. ... and we had to get out." This, along with the pressure members felt that parents were trying to "rescue" children who had joined CoG, encouraged members to "[migrate] abroad -- first to Europe, eventually to Latin America and East Asia". By 1972, COG stated it had 130 communities around the world, and by the mid-1970s, it had "colonies" in an estimated 70 countries. BBC reported 10,000 full-time COG members in the 1970s. In 1976, Berg had introduced a new proselytizing method called Flirty Fishing (or FFing), which encouraged female members to "show God's love" through sexual relationships with potential converts. Flirty Fishing was practiced by members of Berg's inner circle starting in 1973, and was introduced to the general membership in 1976. The Family of Love (1978–1981) The Children of God was abolished in February 1978, and Berg renamed his group "The Family of Love" In what Berg called the "Re-organization Nationalization Revolution" (or RNR). Berg reorganized the movement, dismissing "more than 300 leading members after hearing unspecified 'reports of serious misconduct and abuse of their positions." Reportedly involved were The Chain's abuse of authority, and disagreements within it about the continued use of Flirty Fishing. The group was also accused of sexually abusing and raping minors within the organization, with considerable evidence to support this claim. One eighth of the total membership left the movement. Those who remained became part of a reorganized movement called the Family of Love, and later, The Family. The majority of the group's beliefs remained the same. The Family of Love era was characterized by international expansion. After 1978 Flirty Fishing "increased drastically" and became common practice within the group. A Mo Letter from 1980 (ML #999 May 1980) for example was headlined "The Devil Hates Sex! --- But God Loves It!". In some areas flirty fishers used escort agencies to meet potential converts. According to TFI "over 100,000 received God's gift of salvation through Jesus, and some chose to live the life of a disciple and missionary" as a result of Flirty Fishing. Researcher Bill Bainbridge obtained data from TFI suggesting that, from 1974 until 1987, members had sexual contact with 223,989 people while practicing Flirty Fishing. The Family (1982–1994) According to the Family's official history, the group had "far fewer common standards of conduct" during The Family of Love stage than it had previously. In the late 1980s the group "tightened its standards" "to ensure that all member communities provide a very wholesome environment for all, particularly the children", and changed its name to "The Family". In March 1989, TF issued a statement that, in "early 1985", an urgent memorandum had been sent to all members "reminding them that any such activities [adult–child sexual contact] are within our group" (emphasis in original), and such activities were grounds for immediate excommunication from the group. In January 2005, Claire Borowik, a spokesperson for TFI, stated: Due to the fact that our current zero-tolerance policy regarding sexual interaction between adults and underage minors was not in our literature published before 1986, we came to the realization that during a transitional stage of our movement, from 1978 until 1986, there were cases when some minors were subject to sexually inappropriate advances ... This was corrected officially in 1986, when any contact between an adult and minor (any person under 21 years of age) was declared an excommunicable offense. In the early 1990s, the group broke "years of virtual silence" and began "inviting reporters and religious scholars" to visit its commune in La Habra, California, where at least a Washington Post journalist (Gustav Niebuhr) found its members to be "a clean-cut bunch, friendly and courteous". At that time The Family claimed to have "about 9,000 members worldwide, with about 750 scattered across the United States". The group emphasized its mainstream Christian opposition to abortion, homosexuality, drugs and drunkenness and its respect for Rev. Billy Graham. The Family (1995–2003) After Berg's death in October 1994, Karen Zerby (known in the group as Mama Maria, Queen Maria, Maria David, or Maria Fontaine) assumed leadership of the group. In February 1995, the group introduced the Love Charter, which defined the rights and responsibilities of Charter Members and Homes. The Charter also included the Fundamental Family Rules, a summary of rules and guidelines from past TF publications which were still in effect. In the 1994–95 British court case, the Rt. Hon. Lord Justice Alan Ward ruled that the group, including some of its top leaders, had in the past engaged in abusive sexual practices involving minors and had also used severe corporal punishment and sequestration of minors. He found that by 1995 TF had abandoned these practices and concluded that they were a safe environment for children. Nevertheless, he did require that the group cease all corporal punishment of children in the United Kingdom and denounce any of Berg's writings that were "responsible for children in TF having been subjected to sexually inappropriate behaviour". The Family International (2004–present) The Love Charter is The Family's set governing document that entails each member's rights, responsibilities and requirements, while the Missionary Member Statutes and Fellow Member Statutes were written for the governance of TFI's Missionary member and Fellow Member circles, respectively. FD Homes were reviewed every six months against a published set of criteria. The Love Charter increased the number of single family homes as well as homes that relied on jobs such as self-employment. Recent teachings TFI's recent teachings are based on beliefs which they term the "new [spiritual] weapons". TFI members believe that they are soldiers in the spiritual war of good versus evil for the souls and hearts of men. Spirit Helpers "Spirit Helpers" include angels, other religious and mythical figures, and departed humans, including celebrities; for example the goddess Aphrodite, the Snowman, Merlin, the Sphinx, Elvis, Marilyn Monroe, Audrey Hepburn, Richard Nixon, and Winston Churchill. The Keys of the Kingdom TFI believes that the Biblical passage "I will give you the keys of the kingdom of heaven, and whatsoever you bind on earth will be bound in heaven, and whatsoever you loose on earth will be loosed in heaven", () refers to an increasing amount of spiritual authority that was given to Peter and the early disciples. According to TFI beliefs, this passage refers to keys that were hidden and unused in the centuries that followed, but were again revealed through Karen Zerby as more power to pray and obtain miracles. TFI members call on the various Keys of the Kingdom for extra effect during prayer. The Keys, like most TFI beliefs, were published in magazines that looked like comic-books in order to make them teachable to children. These beliefs are still generally held and practiced, even after the "reboot" documents of 2010. Loving Jesus "Loving Jesus" is a term TFI members use to describe their intimate, sexual relationship with Jesus. TFI describes its "Loving Jesus" teaching as a radical form of bridal theology. They believe the church of followers is Christ's bride, called to love and serve him with wifely fervor; however, this bridal theology is taken further, encouraging members to imagine Jesus is joining them during sexual intercourse and masturbation. Male members are cautioned to visualize themselves as women, in order to avoid a homosexual relationship with Jesus. Many TFI publications, and spirit messages claimed to be from Jesus himself, elaborate this intimate, sexual relation they believe Jesus desires and needs. TFI imagines itself as his special "bride" in graphic poetry, guided visualizations, artwork, and songs. Some TFI literature is not brought into conservative countries for fear it may be classified at customs as pornography. The literature outlining this view of Jesus and his desire for a sexual relationship with believers was edited for younger teens, then further edited for children. Controversy Second-generation adults (known as "SGAs") are adults born or reared in TFI. Anti-TFI sentiment has been publicly expressed by some who have left the group; examples include sisters Celeste Jones, Kristina Jones, and Juliana Buhring, who wrote a book on their lives in TFI. TFI members are expected to respect legal and civil authorities where they live. Members have typically cooperated with appointed authorities, even during the police and social-service raids of their communities in the early 1990s. Criticism The Family has been criticized by the press and the anti-cult movement. Ex-members have accused the Family's leadership of following "a policy of lying to outsiders," being "steeped in a history of sexual deviance" and even meddling "in Third World politics". The Family replies that it is a victim of "persecution." In 1971, an organization called FREECOG was founded by concerned parents and others, including deprogrammer Ted Patrick to "free" members of the COG from their involvement in the group. At least one individual growing up in the family (Verity Carter) during the Children of God era described being sexually abused "from the age of four by members of the... cult, including her own father". She blames the philosophy of David Berg, who told members that "God was love and love was sex", so that sex should not be limited by age or relationship. Carter also complains of being "repeatedly beaten and whipped for the smallest of transgressions", being denied "music or television or culture," or other "contact with the outside world," so that she had "no idea how the world worked" other than how to manipulate the "systemites" (outsiders), like social workers. Author Don Lattin interviewed numerous members of the Family for his book Jesus Freaks. In a review of his book, Paul Burgarino describes Berg as "drawing from the remnants of hippie life—people with nothing to lose, nowhere to go, and no Christian background" to alert them to deviations in Berg's preaching. One ex-Children of God member, Jerry Golland, describes himself at the time of joining the group as penniless and so depressed that the Children of God scraped him "off the street". Members would "learn to spot, you know... a vulnerable person. We called them sheep", Golland told the Canadian Broadcasting Corporation. Pressure to raise money could also be intense. Ex-member Golland says that members who were good at raising money and distributing the pamphlets were called "Shiners". Those with poor sales were called "Shamers". "If you missed your quota you could not come home for dinner", he said. Notable members (past and present) Joined in adulthood Jeremy Spencer, blues slide guitarist and a founding member of Fleetwood Mac, which he left in 1971 when he joined TFI. Raised in the COG and later left Christopher Owens: musician, of US indie band Girls, was brought up in TFI by his parents. Celeste Jones and Kristina Jones: co-authors, along with Juliana Buhring, of Not Without My Sister, an autobiography detailing extensive abuse they suffered in COG. This book is used by the organization RAINN as a reference for child sexual abuse victims. Juliana Buhring: first woman to bicycle around the world and co-author of Not Without My Sister. Rose McGowan: film actress, described her TFI childhood in interviews with Howard Stern, People magazine and later in her book Brave. River Phoenix, Joaquin Phoenix, Rain Phoenix, and Summer Phoenix: actors, were members of the group (with their sister Liberty Phoenix) from 1972 to 1978. River Phoenix, who died of a drug overdose in 1993, told Details magazine in November 1991 that "they're ruining people's lives." Susan Justice: American pop rock singer-songwriter and guitarist, known best for her debut self-recorded album, The Subway Recordings. Tina Dupuy: American journalist and syndicated columnist. Ricky Rodriguez: subject of the suppressed manual advocating adult-child sexual contact, committed a murder-suicide in 2005, killing one of the women who raised and allegedly sexually abused him, then himself. Lauren Hough: author of Leaving is Not the Hardest Thing, brought up in TFI. Flor Edwards, author, who was raised inside the cult before her parents moved out. Dawn Watson: Brazilian, victim of sexual abuse while living in a TFI community. Taylor Stevens, author, brought up in the cult from age 12 until she left in her twenties with her two children. Bexy Cameron, British child member who left aged 15 and later wrote a book about her experiences. Faith Jones, a lawyer, was raised in the group in Macau before leaving. She wrote about her life in the book Sex Cult Nun. Autobiographical accounts Davis, Deborah (Linda Berg) (1984). THE CHILDREN OF GOD: The Inside Story. Zondervan Books, Grand Rapids, Michigan. . Expose by the founder's eldest daughter who left the cult. . Expose by a senior member who left after 20+ years. Behind the tall, foreboding gates of a commune in Brazil, Daniella Mestyanek Young was raised in the religious cult The Children of God, also known as The Family, as the daughter of high-ranking members. Media featuring the group The Jesus Trip (1971), a documentary by Denis Tuohy that has interviews with Children of God members. Children of God (1994), a 63-minute Channel 4 documentary by John Smithson; detailing the Padilla family and the abuse of their three underage daughters and the death of another. Children of God: Lost and Found, a 75-minute documentary by Noah Thomson, featured at the 2007 Slamdance Film Festival. Cult Killer: The Rick Rodriguez Story (53-minute UK documentary with transcript). In the first episode of Louis Theroux's Weird Weekends, "Born Again Christians", Louis visits a Texas TFI family. Buzzcocks mentions the group (as "Children Of God") in their song, "Orgasm Addict". RedLetterMedia featured the Family International video "S.O.S." on an episode of "Best of the Worst." Mentioned in Tony Robbins: I Am Not Your Guru documentary at 52 minutes of the film as an organization where children are forced to have sex from the age of six. The Parcast Podcast Cults: Episodes 11 and 12. Citizen Rose: A five part documentary series shown on the E! Channel. The first episode premiered on January 30, 2018. The series follows actress Rose McGowan who was born into the cult. The Last Podcast on the Left did a four part series on the cult: Episodes 248-251 The Dan Cummins podcast Timesuck covered the cult in episode 104, "The Children of God Sex Cult." AJJ released a song entitled "Children of God" on their 2014 album Christmas Island. A&E's Cults and Extreme Belief, episode 3 (2018) is about the Children of God. See also Comet Kohoutek was viewed by David Berg as a prophetic sign of imminent disaster. Jim Palosaari co-formed the Jesus People Army, left it before the group joined the Children of God, and tried to convince Linda Meissner not to join it. Love bombing describes a manipulative style of recruiting. Panton Hill, Victoria is the location of one of the communes, where a large government raid occurred and many children were removed by social services. References Further reading Academic Chancellor, James (2000). Life in The Family: An Oral History of the Children of God. University of Syracuse Press, Syracuse, NY. Bainbridge, William Sims (2002). The Endtime Family: Children of God. State University of New York Press. . Bainbridge, William Sims (1996). The Sociology of Religious Movements. Routledge. . Barker, Eileen. (1989). New Religious Movements, A Practical Introduction. Her Majesty's Stationery Office. . Barker, Eileen. (2021). "Children of God/The Family International Armageddon" . In James Crossley and Alastair Lockhart (eds.) Critical Dictionary of Apocalyptic and Millenarian Movements. Barrett, DV (1996). Sects, Cults and Alternative Religions. Blandford A. Cassell. . Lewis, James R, and Melton, J. Gordon (eds). (1994). Sex, Slander, and Salvation: Investigating The Family/Children of God. Center for Academic Press, Stanford, CA. Lynch, Dalva, and Paul Carden (1990). "Inside the 'Heavenly Elite': The Children of God Today." . Christian Research Journal, pp 16. McFarland, Robert (1994). "The Children of God." The Journal of Psychohistory 4(21). Melton, J. Gordon (2004). The Children of God, "The Family" (Studies in Contemporary Religion vol. 7). Signature Books. . Melton, J. Gordon (2004). The Family International Britannica Article Melton, J. Gordon and Robert L. Moore (1982). The Cult Experience: Responding to the New Religious Pluralism. The Pilgrim Press, New York, USA. Palmer, Susan J. (1994). "Heaven's Children: The Children of God's Second Generation" in Sex, Slander, and Salvation, op. cit. Palmer, Susan J., and Charlotte Hardman eds. (1999). Children in New Religions (3rd ed.). Rutgers University Press. . Shepherd, Gary, and Lawrence Lilliston (1994). "Field Observations of Young People's Experience and Role in The Family" in Sex, Slander, and Salvation, op. cit. Shepherd, Gary, and Shepherd, Gordon (August 2005). "Accommodation and Reformation in The Family/Children of God" , Nova Religio (Journal of the University of California) Shepherd, Gary and Shepherd, Gordon (Spring 2000)."The Moral Career of a New Religious Movement" The Oakland Journal. Wilson, Bryan and Jamie Cresswell, eds. (1999). New Religious Movements: Challenge and Response. Routledge, London, UK. Wright, Stuart (1987). Leaving Cults: The Dynamics of Defection. Society for the Scientific Study of Religion. Washington, D.C., USA. (Contains interviews with ex-members of three groups, among others the Children of God) Van Zandt, David (1991). Living in the Children of God. Princeton University Press, Princeton, New Jersey. Young, Shawn David, Hippies, Jesus Freaks, and Music (Ann Arbor: Xanedu/Copley Original Works, 2005). . Journalistic and popular McManus, Una (1980). Not for a Million Dollars. Impact Books. . Williams, Miriam (1999). Heaven's Harlots: My Fifteen Years As a Sacred Prostitute in the Children of God Cult. Quill. . "30 Members of Children of God arrested" (September 2, 1993). Washington Post, pp. A05 "The Family" and Final Harvest" (June 2, 1993). Washington Post, pp. A01 Goodstein, Laurie (2005), "Murder and Suicide Reviving Claims of Child Abuse in Cult ", The New York Times, January 15, 2005, pg. A-1 Don Lattin: Jesus Freaks: A True Story of Murder and Madness on the Evangelical Edge. HarperOne. . Mahoney, Mary (2020). Abnormal Normal: My Life in the Children of God'' Jones, Faith (2021). "Sex Cult Nun". Harper Collins. External links Official DavidBerg.org – Official website explaining David Brandt Berg's mission, vision and message. KarenZerby.org – Karen Zerby's official site. TFICharter.com – Official Governing Documents of The Family International. Children of God.com – Official history of the COG (pre-TFI). Other xFamily – Wiki detailing TFI; includes large collections of multimedia, press coverage, and internal TFI publications. xFamily PubsDB – a near-complete database of all writings by David Berg and Karen Zerby. exfamily.org – information, forums, links, etc. about TFI by former first-generation members. The Family International Cults 1968 establishments in California Christian new religious movements Christian organizations established in the 20th century Jesus movement Religious organizations established in 1968 Religious belief systems founded in the United States
7768
https://en.wikipedia.org/wiki/Crete%20Senesi
Crete Senesi
The Crete Senesi refers to an area of the Italian region of Tuscany immediately to the south of Siena. It consists of a range of hills and woods among villages and includes the comuni of Asciano, Buonconvento, Monteroni d'Arbia, Rapolano Terme and San Giovanni d'Asso, all within the province of Siena. They border to the north with the Chianti Senese area, to the east with Val di Chiana and to the south-west with Val d'Orcia. Nearby is also the semi-arid area known as the Accona Desert. Crete Senesi are literally the "clays of Siena": the distinctive grey colouration of the soil gives the landscape an appearance often described as lunar. This characteristic clay, known as mattaione, represents the sediments of the Pliocene sea which covered the area between 2.5 and 4.5 million years ago. The landscape is characterized by barren and gently undulating hills, solitary oaks and cypresses, isolated farms at the top of the heights, stretches of wood and ponds of rainwater (commonly referred as fontoni, literally "big springs") in the valleys. Badlands and are typical conformations of the land. Perhaps the most notable edifice of this area is the Abbey of Monte Oliveto Maggiore, located 10 km south of Asciano. The region is known for its production of white truffles, and hosts a festival and a museum dedicated to the rare fungus (genus Tuber). Gallery References External links Official website Terre di Siena: Crete Senesi Videos of the Crete Senesi Crete Senesi Landscape Geography of Tuscany Province of Siena
7957
https://en.wikipedia.org/wiki/Kennedy%20family
Kennedy family
The Kennedy family is an American political family that has long been prominent in American politics, public service, entertainment, and business. In 1884, 35 years after the family's arrival from Ireland, Patrick Joseph "P. J." Kennedy became the first Kennedy elected to public office, serving in the Massachusetts state legislature until 1895. At least one Kennedy family member served in federal elective office from 1947, when P. J. Kennedy's grandson John F. Kennedy became a member of Congress from Massachusetts, until 2011, when Patrick J. Kennedy II (John's nephew) retired as a member of the U.S. House of Representatives from Rhode Island. P. J.'s son Joseph P. Kennedy Sr. had nine children with his wife Rose Fitzgerald Kennedy, including John F. Kennedy, who served in both houses of the United States Congress and as U.S. President; Robert F. Kennedy, who served as U.S. Attorney General and as a U.S. Senator; and Ted Kennedy, who served more than 46 years in the U.S. Senate. Other descendants include members of the U.S. House of Representatives, two U.S. ambassadors, one U.S. envoy, a lieutenant governor, three state legislators (one of whom also served in the U.S. House of Representatives), and one mayor. Joseph and Rose's daughter, Eunice, founded the National Institute of Child Health and Human Development (part of the National Institutes of Health), and the Special Olympics. Eunice's daughter, Maria Shriver, served as First Lady of California. Other descendants of Joseph and Rose Kennedy have been active as lawyers, authors, and activists on behalf of those with physical and intellectual disabilities. History According to genealogist Brian Kennedy in his work JFK's Irish O'Kennedy Ancestors, the Kennedys who would go on to play a significant role in the United States of America, originated from an Irish clan called Ó Cinnéide Fionn (one of the three Irish Gaelic Ó Cinnéide clans who ruled the kingdom of Ormond, along with the Ó Cinnéide Donn and Ó Cinnéide Ruadh). Their progenitor, Diarmaid Ó Cinnéide Fionn, held Knigh Castle close to what is today Puckane, County Tipperary in 1546. From there, having lost out to the New English order in the Kingdom of Ireland, they ended up in Dunganstown, New Ross, County Wexford by 1740. Patrick Kennedy was born there in 1823. Patrick Kennedy (1823–1858) and Bridget Murphy (1824–1888) sailed from Ireland to East Boston in 1849. Patrick worked in East Boston as a barrel maker, or cooper, and had five children with Bridget. Their youngest, Patrick Joseph "P. J." Kennedy, went into business and served in the Massachusetts state legislature from 1884 to 1895. P. J. and Mary Augusta Hickey were the parents of four children. Their oldest was Joseph Patrick "Joe" Kennedy Sr., a businessman who amassed a private fortune in banking and securities trading, to which he further expanded by investing in filmmaking and real estate. He also founded Somerset Importers and owned the Merchandise Mart in Chicago, Illinois. In 1914, Joseph Sr. married Rose Fitzgerald, the eldest daughter of John F. "Honey Fitz" Fitzgerald, who served as the mayor of Boston (1906–08, 1910–14). The couple would have nine children: Joseph Jr., John (called Jack), Rose Marie (called Rosemary), Kathleen, Eunice, Patricia, Robert (called Bobby), Jean, and Edward (called Ted). Joseph Sr. was appointed by President Franklin D. Roosevelt as the first chairman of the Securities and Exchange Commission (SEC), chairman of the Maritime Commission, and U.S. ambassador to the United Kingdom from 1938 to 1940. He served on The Hoover Commission, officially named the Commission on Organization of the Executive Branch of the Government, from 1947 to 1949; the commission was appointed by President Harry Truman to recommend administrative changes in the federal government. His wife, Rose, was appointed Papal Countess of the Holy Roman Church by Pope Pius XII. Continued public service Every Kennedy elected to public office has served as a Democrat, while other members of the family have worked for the party or held Cabinet posts in Democratic administrations. Many have attended Harvard University, and the family has contributed greatly to that university's John F. Kennedy School of Government. Joseph Sr. expected his eldest son, Joseph Jr., to go into politics and to ultimately be elected president. Joseph Jr. was elected as a Massachusetts delegate to the 1940 Democratic National Convention and enlisted in the U.S. Navy after the United States entered World War II. He was killed in 1944 when the bomber he was piloting exploded in flight. Joseph Sr.'s desire to see the family involved in politics and government then focused on John, who had considered a career as a journalist, having authored a book (Why England Slept) and done some reporting for Hearst Newspapers. After returning from Navy service, John served in the U.S. House of Representatives representing Massachusetts's 11th congressional district from 1947 to 1953, and then as U.S. Senator from Massachusetts from 1953 to 1960. In the 1960 presidential election, John narrowly defeated Republican opponent Richard Nixon. During John's administration, Robert served as attorney general; his brother-in-law Sargent Shriver served as director of the new Peace Corps, and Ted was elected to the U.S. Senate from Massachusetts, occupying his brother's former Senate seat until his death in 2009. The Kennedy administration's accomplishments include the Alliance for Progress with Latin America, the establishment of the Peace Corps, a peaceful resolution to the Cuban Missile Crisis in October 1962, the Nuclear Test Ban Treaty of 1963, the 24th Amendment ending the poll tax, the continuation of the Apollo program with the goal of landing a man on the Moon, and the introduction of the Civil Rights Act of 1964 to Congress (signed into law by Kennedy's successor Lyndon B. Johnson). The family was the subject of intense media coverage during and after John Kennedy's presidency. Ted served in the Senate with his brother Robert (1965–1968), and was serving in the Senate when his nephew, Joseph P. II (1987–1999), and son, Patrick J. (1995–2011), served in the U.S. House of Representatives representing Massachusetts's 8th congressional district and Rhode Island's 1st congressional district respectively. In November 2012, Joseph P. Kennedy III, son of former Rep. Joseph P. Kennedy II and grandson of the late Sen. Robert F. Kennedy, was elected to the U.S. House of Representatives from Massachusetts's 4th congressional district. In 2020, he lost the U.S. Senate primary election in Massachusetts to incumbent Ed Markey, the first Kennedy to ever lose an election in the state. In the 2020s, three Kennedy family members were serving as U.S. ambassadors or envoys. Victoria Reggie Kennedy, second wife of Ted Kennedy, was named in 2021 by President Biden as U.S. ambassador to Austria. Caroline Kennedy, daughter of President Kennedy, was named in 2022 by President Biden as U.S. ambassador to Australia; she previously served as U.S. ambassador to Japan under President Barack Obama. In the same year, Joseph P. Kennedy III was named by President Biden as U.S. special envoy to Northern Ireland. Family incidents Family incidents led Senator Ted Kennedy to wonder, in a televised statement about the Chappaquiddick incident in 1969, whether "some awful curse did actually hang over all the Kennedys." Some of the events endured by the Kennedy clan include: In 1941, Rosemary underwent a non-consensual lobotomy intended to correct her violent mood swings, convulsions, and intellectual disability. The operation left her incapacitated for the rest of her life. Joseph Jr. died in 1944 when the Navy bomber he was piloting exploded in mid-flight. Kathleen died in a plane crash in France in 1948. John and Robert were assassinated, in 1963 and 1968 respectively. In 1964, Ted was nearly killed when his plane crashed in an apple orchard near Southampton, Massachusetts. (Legislative aide Edward Moss and the pilot were killed in the crash.) Ted was seriously injured and spent months in a hospital recovering from a severe back injury, a punctured lung, broken ribs and internal bleeding. In later generations, Robert's son David died of a drug overdose in 1984; another son, Michael, died from injuries sustained in a skiing accident in 1997. John's son, John Jr., died in a plane crash (along with his wife Carolyn and sister-in-law Lauren) off the coast of Martha's Vineyard in 1999. Ted's daughter Kara and Pat's son Christopher died of heart attacks, in 2011 and 2018 respectively. Robert's granddaughter, Saoirse Kennedy Hill, died of a drug overdose in 2019 aged 22. In April 2020, another granddaughter, Maeve Kennedy McKean, a senior advisor for human rights in the Obama administration, and her eight-year-old son, Gideon Joseph Kennedy McKean, disappeared in Chesapeake Bay after embarking in a canoe to retrieve a ball. Maeve McKean's body was recovered the following week, and her son's two days later, about 2.5 miles from her mother's home on the Chesapeake Bay. Government offices held Patrick Joseph Kennedy: Massachusetts state Representative 1884–1889; Massachusetts state Senator, 1889–1895. Joseph Patrick Kennedy Sr.: Chairman of the U.S. Securities and Exchange Commission, 1934–1935; chairman of the United States Maritime Commission, 1936–1938; United States Ambassador to the United Kingdom, 1938–1940. John Fitzgerald Kennedy: United States Representative from Massachusetts, 1947–1953; United States Senator from Massachusetts, 1953–1960; President of the United States, 1961–1963. Caroline Kennedy: United States Ambassador to Japan, 2013–2017; United States Ambassador to Australia 2022-. Eunice Kennedy Shriver Bobby Shriver: Santa Monica, California City Council member, 2004–2012; Mayor of Santa Monica, 2010. Mark Kennedy Shriver: Maryland state Delegate, 1995–2003. Robert Francis Kennedy: United States Attorney General 1961–1964; United States Senator from New York, 1965–1968. Kathleen Kennedy Townsend: Lieutenant governor of Maryland, 1995–2003. Joseph P. Kennedy II: United States Representative from Massachusetts, 1987–1999. Joseph P. Kennedy III: United States Representative from Massachusetts, 2013–2021; U.S. envoy to Northern Ireland, 2022-. Jean Kennedy Smith: United States Ambassador to Ireland, 1993–1998. Edward Moore Kennedy: United States Senator from Massachusetts, 1962–2009. Edward M. Kennedy Jr.: Connecticut state Senator, 2015–2019. Patrick J. Kennedy: Rhode Island state Representative, 1989–1993; United States Representative from Rhode Island, 1995–2011. Other Kennedy relatives and spouses have served in government: Andrew Cuomo (then-husband of Kerry Kennedy): United States Secretary of Housing and Urban Development, 1997-2001. He later served as New York state attorney general and governor. John F. Fitzgerald (maternal grandfather of John, Robert and Edward Kennedy): Massachusetts state senator, 1892-94; member, U.S. House of Representatives, 1895-1901, 1919; mayor of Boston, Massachusetts, 1906-08, 1910-14. Charles Hickey (uncle of Joseph P. Kennedy, granduncle of John, Robert and Edward Kennedy): mayor of Brockton, Massachusetts, 1913. Victoria Reggie Kennedy (widow of Ted Kennedy): U.S. ambassador to Austria, 2022–present Arnold Schwarzenegger (then-husband of Maria Shriver): governor of California, 2003–2011 Sargent Shriver (husband of Eunice Kennedy Shriver): president of the Chicago Board of Education, 1955–1960; director of the Peace Corps, 1961–1966; director of the Office of Economic Opportunity, 1964–1968; U.S. ambassador to France, 1968–1970 There was a member of the Kennedy family in public office nearly continuously from 1946, when John F. Kennedy was elected to the U.S. House of Representatives, until early 2011, when Patrick J. Kennedy left the House. The only exception was the period between John F. Kennedy's resignation from the Senate on December 22, 1960, and his assumption of the office of President on January 20, 1961. In 2013, two years after Patrick Kennedy left the House, Joseph P. Kennedy III was elected U.S. Representative from Massachusetts and served until 2021. Below is a timeline of the Kennedys' tenure in the U.S. Congress. Timeline Heraldry In 1961, John F. Kennedy was presented with a grant of arms for all the descendants of Patrick Kennedy (1823–1858) from the Chief Herald of Ireland. The design of the arms (three gold closed helmets on a black field) strongly alludes to symbols in the coats of arms of the O'Kennedys of Ormonde and the FitzGeralds of Desmond, from whom the family is descended. The crest is an armored hand holding four arrows between two olive branches, elements taken from the coat of arms of the United States of America and also symbolic of Kennedy and his brothers. See also List of Kennedy family members References Bibliography Haas, Lawrence J. The Kennedys in the World: How Jack, Bobby, and Ted Remade America's Empire (2021) excerpt Hunt, Amber, and David Batcher. Kennedy Wives: Triumph and Tragedy in America's Most Public Family (2014) excerpt Kessler, Ronald. The sins of the father: Joseph P. Kennedy and the dynasty he founded (St. Martin's Press, 1996). Klein, Edward. The Kennedy Curse: Why tragedy has haunted America's first family for 150 years (Macmillan, 2003). Leamer, Laurence. The Kennedy women: The saga of an American family (Ballantine Books, 1996). excerpt Leamer, Laurence. The Kennedy Men: 1901-1963 (2001) excerpt Leamer, Laurence. Sons of Camelot: The Fate of an American Dynasty (2005) excerpt Nasaw, David. The Patriarch: The Remarkable Life and Turbulent Times of Joseph P. Kennedy (2012); scholarly biography. External links The Kennedys: A Family Tree, St. Petersburg Times Kennedy Family Tree, The New York Times Kennedy Family, The Political Graveyard American families of Irish ancestry Business families of the United States First Families of the United States Roman Catholic families County Wexford Irish-American culture Irish-American history Massachusetts Democrats People from Barnstable, Massachusetts People from Boston People from Brookline, Massachusetts
8090
https://en.wikipedia.org/wiki/Day%20of%20the%20Tentacle
Day of the Tentacle
Day of the Tentacle, also known as Maniac Mansion II: Day of the Tentacle, is a 1993 graphic adventure game developed and published by LucasArts. It is the sequel to the 1987 game Maniac Mansion. The plot follows Bernard Bernoulli and his friends Hoagie and Laverne as they attempt to stop the evil Purple Tentacle - a sentient, disembodied tentacle - from taking over the world. The player takes control of the trio and solves puzzles while using time travel to explore different periods of history. Dave Grossman and Tim Schafer co-led the game's development, their first time in such a role. The pair carried over a limited number of elements from Maniac Mansion and forwent the character selection aspect to simplify development. Inspirations included Chuck Jones cartoons and the history of the United States. Day of the Tentacle was the eighth LucasArts game to use the SCUMM engine. The game was released simultaneously on floppy disk and CD-ROM to critical acclaim and commercial success. Critics focused on its cartoon-style visuals and comedic elements. Day of the Tentacle has featured regularly in lists of "top" games published more than two decades after its release, and has been referenced in popular culture. A remastered version of Day of the Tentacle was developed by Schafer's current studio, Double Fine Productions, and released in March 2016, for Windows, OS X, Linux, PlayStation 4 and PlayStation Vita with an Xbox One release in October 2020. Gameplay Day of the Tentacle follows the point-and-click two-dimensional adventure game formula, first established by the original Maniac Mansion. Players direct the controllable characters around the game world by clicking with the computer mouse. To interact with the game world, players choose from a set of nine commands arrayed on the screen (such as "pick up", "use", or "talk to") and then on an object in the world. This was the last SCUMM game to use the original interface of having the bottom of the screen being taken up by a verb selection and inventory; starting with the next game to use the SCUMM engine, Sam & Max Hit the Road, the engine was modified to scroll through a more concise list of verbs with the right mouse button and having the inventory on a separate screen. Day of the Tentacle uses time travel extensively; early in the game, the three main protagonists are separated across time by the effects of a faulty time machine. The player, after completing certain puzzles, can then freely switch between these characters, interacting with the game's world in separate time periods. Certain small inventory items can be shared by placing the item into the "Chron-o-Johns", modified portable toilets that instantly transport objects to one of the other time periods, while other items are shared by simply leaving the item in a past time period to be picked up by a character in a future period. Changes made to a past time period will affect a future one, and many of the game's puzzles are based on the effect of time travel, the aging of certain items, and alterations of the time stream. For example, one puzzle requires the player, while in the future era where Purple Tentacle has succeeded, to send a medical chart of a Tentacle back to the past, having it used as the design of the American flag, then collecting one such flag in the future to be used as a Tentacle disguise to allow that character to roam freely. The whole original Maniac Mansion game can be played on a computer resembling a Commodore 64 inside the Day of the Tentacle game; this practice has since been repeated by other game developers, but at the time of Day of the Tentacle's release, it was unprecedented. Plot Five years after the events of Maniac Mansion, Purple Tentacle—a mutant monster and lab assistant created by mad scientist Dr. Fred Edison—drinks toxic sludge from a river behind Dr. Fred's laboratory. The sludge causes him to grow a pair of flipper-like arms, develop vastly increased intelligence, and have a thirst for global domination. Dr. Fred plans to resolve the issue by killing Purple Tentacle and his harmless, friendly brother Green Tentacle, but Green Tentacle sends a plea of help to his old friend, the nerd Bernard Bernoulli. Bernard travels to the Edison family motel with his two housemates, deranged medical student Laverne and roadie Hoagie, and frees the tentacles. Purple Tentacle escapes to resume his quest to take over the world. Since Purple Tentacle's plans are flawless and unstoppable, Dr. Fred decides to use his Chron-o-John time machines to send Bernard, Laverne, and Hoagie to the day before to turn off his Sludge-o-Matic machine, thereby preventing Purple Tentacle's exposure to the sludge. However, because Dr. Fred used an imitation diamond rather than a real diamond as a power source for the time machine, the Chron-o-Johns break down in operation. Laverne is sent 200 years in the future, where humanity has been enslaved and Purple Tentacle rules the world from the Edison mansion, while Hoagie is dropped 200 years in the past, where the motel is being used by the Founding Fathers as a retreat to write the United States Constitution. Bernard is returned to the present. To salvage Dr. Fred's plan, Bernard must acquire a replacement diamond for the time machine, while both Hoagie and Laverne must restore power to their respective Chron-o-John pods by plugging them in. To overcome the lack of electricity in the past, Hoagie recruits the help of Benjamin Franklin and Dr. Fred's ancestor, Red Edison, to build a superbattery to power his pod, while Laverne evades capture by the tentacles long enough to run an extension cord to her unit. The three send small objects back and forth in time through the Chron-o-Johns and make changes to history to help the others complete their tasks. Eventually, Bernard uses Dr. Fred's family fortune of royalties from the Maniac Mansion TV series to purchase a real diamond, while his friends manage to power their Chron-o-Johns. Soon, the three are reunited in the present. Purple Tentacle arrives, hijacks a Chron-o-John, and takes it to the previous day to prevent them from turning off the sludge machine; he is pursued by Green Tentacle in another pod. With only one Chron-o-John pod left, Bernard, Hoagie, and Laverne use it to pursue the tentacles to the previous day, while Dr. Fred uselessly tries to warn them of using the pod together, referencing the film The Fly. Upon arriving, the trio exit the pod only to discover that they have been turned into a three-headed monster, their bodies merging into one during the transfer. Meanwhile, Purple Tentacle has used the time machine to bring countless versions of himself from different moments in time to the same day to prevent the Sludge-o-Matic from being deactivated. Bernard and his friends defeat the Purple Tentacles guarding the Sludge-o-Matic, turn off the machine, and prevent the whole series of events from ever happening. Returning to the present, Dr. Fred discovers that the three have not been turned into a monster at all but have just gotten stuck in the same set of clothes; they are then ordered by Dr. Fred to get out of his house. The game ends with the credits rolling over a tentacle-shaped American flag, one of the more significant results of their tampering in history. Development Following a string of successful adventure games, LucasArts assigned Dave Grossman and Tim Schafer to lead development of a new game. The two had previously assisted Ron Gilbert with the creation of The Secret of Monkey Island and Monkey Island 2: LeChuck's Revenge, and the studio felt that Grossman and Schafer were ready to manage a project. The company believed that the pair's humor matched well with that of Maniac Mansion and suggested working on a sequel. The two developers agreed and commenced production. Gilbert and Gary Winnick, the creators of Maniac Mansion, collaborated with Grossman and Schafer on the initial planning and writing. The total budget for the game was about $600,000, according to Schafer. Creative design In planning the plot, the four designers considered a number of concepts, eventually choosing an idea of Gilbert's about time travel that they believed was the most interesting. The four discussed what time periods to focus on, settling on the Revolutionary War and the future. The Revolutionary War offered opportunities to craft many puzzles around that period, such as changing the Constitution to affect the future. Grossman noted the appeal of the need to make wide-sweeping changes such as the Constitution just to achieve a small personal goal, believing this captured the essence of adventure games. The future period allowed them to explore the nature of cause and effect without any historical bounds. Grossman and Schafer decided to carry over previous characters that they felt were the most entertaining. The two considered the Edison family "essential" and chose Bernard because of his "unqualified nerdiness". Bernard was considered "everyone's favorite character" from Maniac Mansion, and was the clear first choice for the protagonists. The game's other protagonists, Laverne and Hoagie, were based on a Mexican ex-girlfriend of Grossman's and a Megadeth roadie named Tony that Schafer had met, respectively. Schafer and Grossman planned to use a character selection system similar to the first game but felt that it would have complicated the design process and increased production costs. Believing that it added little to the gameplay, they removed it early in the process and reduced the number of player characters from six to three. The dropped characters included Razor, a female musician from the previous game; Moonglow, a short character in baggy clothes; and Chester, a black beat poet. Ideas for Chester, however, morphed into new twin characters in the Edison family. The smaller number of characters reduced the strain on the game's engine in terms of scripting and animation. The staff collaboratively designed the characters. They first discussed the character personalities, which Larry Ahern used to create concept art. Ahern wanted to make sure that the art style was consistent and the character designs were established early, in contrast to what had happened with Monkey Island 2, in which various artists came in later to help fill in necessary art assets as necessary, creating a disjointed style. Looney Tunes animation shorts, particularly the Chuck Jones-directed Rabbit of Seville, What's Opera, Doc? and Duck Dodgers in the 24½th Century inspired the artistic design. The cartoonish style also lent itself to providing larger visible faces to enable more expressive characters. Peter Chan designed backgrounds, spending around two days to progress from concept sketch to final art for each background. Chan too used Looney Tunes as influence for the backgrounds, trying to emulate the style of Jones and Maurice Noble. Ahern and Chan went back and forth with character and background art to make sure both styles worked together without too much distraction. They further had Jones visit their studio during development to provide input into their developing art. The choice of art style inspired further ideas from the designers. Grossman cited cartoons featuring Pepé Le Pew, and commented that the gag involving a painted white stripe on Penelope Pussycat inspired a puzzle in the game. The artists spent a year creating the in-game animations. The script was written in the evening when fewer people were in the office. Grossman considered it the easiest aspect of production, but encountered difficulties when writing with others around. Grossman and Schafer brainstormed regularly to devise the time travel puzzles and collaborated with members of the development team as well as other LucasArts employees. They would identify puzzle problems and work towards a solution similar to how the game plays. Most issues were addressed prior to programming, but some details were left unfinished to work on later. The staff conceived puzzles involving the U.S.'s early history based on their memory of their compulsory education, and using the more legendary aspects of history, such as George Washington cutting down a cherry tree to appeal to international audiences. To complete the elements, Grossman researched the period to maintain historical accuracy, visiting libraries and contacting reference librarians. The studio, however, took creative license towards facts to fit them into the game's design. Day of the Tentacle features a four-minute-long animated opening credit sequence, the first LucasArts game to have such. Ahern noted that their previous games would run the credits over primarily still shots which would only last for a few minutes, but with Tentacle, the team had grown so large that they worried this approach would be boring to players. They assigned Kyle Balda, an intern at CalArts, to create the animated sequence, with Chan helping to create minimalist backgrounds to aid in the animation. Originally this sequence was around seven minutes long, and included the three characters arriving at the mansion and releasing Purple Tentacle. Another LucasArts designer, Hal Barwood, suggested they cut it in half, leading to the shortened version as in the released game, and having the player take over when they arrive at the mansion. Technology and audio Day of the Tentacle uses the SCUMM engine developed for Maniac Mansion. LucasArts had gradually modified the engine since its creation. For example, the number of input verbs was reduced and items in the character's inventory are represented by icons rather than text. While implementing an animation, the designers encountered a problem later discovered to be a limitation of the engine. Upon learning of the limitation, Gilbert reminisced about the file size of the first game. The staff then resolved to include it in the sequel. Day of the Tentacle was the first LucasArts adventure game to feature voice work on release. The game was not originally planned to include voice work, as at the time, the install base for CD-ROM was too low. As they neared the end of 1992, CD-ROM sales grew significantly. The general manager of LucasArts, Kelly Flock, recognizing that the game would not be done in time by the end of the year to make the holiday release, suggested that the team include voice work for the game, giving them more time. Voice director Tamlynn Barra managed that aspect of the game. Schafer and Grossman described how they imagined the characters' voices and Barra sought audition tapes of voice actors to meet the criteria. She presented the best auditions to the pair. Schafer's sister Ginny was among the auditions, and she was chosen for Nurse Edna. Schafer opted out of the decision for her selection to avoid nepotism. Grossman and Schafer encountered difficulty selecting a voice for Bernard. To aid the process, Grossman commented that the character should sound like Les Nessman from the television show WKRP in Cincinnati. Barra responded that she knew the agent of the character's actor, Richard Sanders, and brought Sanders on the project. Denny Delk and Nick Jameson were among those hired, and provided voice work for around five characters each. Recording for the 4,500 lines of dialog occurred at Studio 222 in Hollywood. Barra directed the voice actors separately from a sound production booth. She provided context for each line and described aspects of the game to aid the actors. The voice work in Day of the Tentacle was widely praised for its quality and professionalism in comparison to Sierra's talkie games of the period which suffered from poor audio quality and limited voice acting (some of which consisted of Sierra employees rather than professional talent). The game's music was composed by Peter McConnell, Michael Land and Clint Bajakian. The three had worked together to share the duties equally of composing the music for Monkey Island 2 and Fate of Atlantis, and continued this approach for Day of the Tentacle. According to McConnell, he had composed most of the music taking place in the game's present, Land for the future, and Bajakian for the past, outside of Dr. Fred's theme for the past which McConnell had done. The music was composed around the cartoonish nature of the gameplay, further drawing on Looney Tunes' use of parodying classical works of music, and playing on set themes for all of the major characters in the game. Many of these themes had to be composed to take into account different processing speeds of computers at the time, managed by the iMUSE music interface; such themes would include shorter repeating patterns that would play while the game's screen scrolled across, and then once the screen was at the proper place, the music would continue on to a more dramatic phrase. Day of the Tentacle was one of the first games concurrently released on CD-ROM and floppy disk. A floppy disk version was created to accommodate consumers that had yet to purchase CD-ROM drives. The CD-ROM format afforded the addition of audible dialog. The capacity difference between the two formats necessitated alterations to the floppy disk version. Grossman spent several weeks reducing files sizes and removing files such as the audio dialog to fit the game onto six diskettes. Reception Day of the Tentacle was a moderate commercial success; according to Edge, it sold roughly 80,000 copies by 2009. Tim Schafer saw this as an improvement over his earlier projects, the Monkey Island games, which had been commercial flops. The game was critically acclaimed. Charles Ardai of Computer Gaming World wrote in September 1993: "Calling Day of the Tentacle a sequel to Maniac Mansion ... is a little like calling the space shuttle a sequel to the slingshot". He enjoyed the game's humor and interface, and praised the designers for removing "dead end" scenarios and player character death. Ardai lauded the voice acting, writing that it "would have done the late Mel Blanc proud", and compared the game's humor, animation, and camera angles to "Looney Toons gems from the 40s and 50s". He concluded: "I expect that this game will keep entertaining people for quite some time to come". In April 1994 the magazine said of the CD version that Sanders's Bernard was among "many other inspired performances", concluding that "Chuck Jones would be proud". In May 1994 the magazine said of one multimedia kit bundling the CD version that "it packs more value into the kit than the entire software packages of some of its competitors". Sandy Petersen of Dragon stated that its graphics "are in a stupendous cartoony style", while praising its humor and describing its sound and music as "excellent". Although the reviewer considered it "one of the best" graphic adventure games, he noted that, like LucasArts' earlier Loom, it was extremely short; he wrote that he "felt cheated somehow when I finished the game". He ended the review, "Go, Lucasfilm! Do this again, but do make the next game longer!". Phil LaRose of The Advocate called it "light-years ahead of the original", and believed that its "improved controls, sound and graphics are an evolutionary leap to a more enjoyable gaming experience". He praised the interface, and summarized the game as "another of the excellent LucasArts programs that place a higher premium on the quality of entertainment and less on the technical knowledge needed to make it run". The Boston Heralds Geoff Smith noted that "the animation of the cartoonlike characters is of TV quality", and praised the removal of dead ends and character death. He ended: "It's full of lunacy, but for anyone who likes light-hearted adventure games, it's well worth trying". Vox Day of The Blade called its visuals "well done" and compared them to those of The Ren & Stimpy Show. The writer praised the game's humor, and said that "both the music and sound effects are hilarious"; he cited the voice performance of Richard Sanders as a high point. He summarized the game as "both a good adventure and a funny cartoon". Lim Choon Wee of the New Straits Times highly praised the game's humor, which he called "brilliantly funny". The writer commented that the game's puzzles relied on "trial and error" with "no underlying logic", but opined that the game "remains fun" despite this issue, and concluded that Day of the Tentacle was "definitely the comedy game of the year". Daniel Baum of The Jerusalem Post called it "one of the funniest, most entertaining and best-programmed computer games I have ever seen", and lauded its animation. He wrote that the game provided "a more polished impression" than either The Secret of Monkey Island or Monkey Island 2: LeChuck's Revenge. The writer claimed that its high system requirements were its only drawback, and believed that a Sound Blaster card was required to fully appreciate the game. In a retrospective review, Adventure Gamers' Chris Remo wrote: "If someone were to ask for a few examples of games that exemplify the best of the graphic adventure genre, Day of the Tentacle would certainly be near the top". Day of the Tentacle has been featured regularly in lists of "top" games. In 1994, PC Gamer US named Day of the Tentacle the 46th best computer game ever. In June 1994 it and Gabriel Knight: Sins of the Fathers won Computer Gaming Worlds Adventure Game of the Year award. The editors wrote that "Day of the Tentacles fluid animation sequences underscore a strong script and solid game play ... story won out over technological innovation in this genre". In 1996, the magazine ranked it as the 34th best game of all time, writing: "DOTT completely blew away its ancestor, Maniac Mansion, with its smooth animated sequences, nifty plot and great voiceovers". Adventure Gamers included the game as the top entry on its 20 Greatest Adventure Games of All Time List in 2004, and placed it sixth on its Top 100 All-Time Adventure Games in 2011. The game has appeared on several IGN lists. The website rated it number 60 and 84 on its top 100 games list in 2005 and 2007, respectively. IGN named Day of the Tentacle as part of their top 10 LucasArts adventure games in 2009, and ranked the Purple Tentacle 82nd in a list of top 100 videogame villains in 2010. ComputerAndVideoGames.com ranked it at number 30 in 2008, and GameSpot also listed Day of the Tentacle as one of the greatest games of all time. Legacy Fans of Day of the Tentacle created a webcomic, The Day After the Day of the Tentacle, using the game's graphics. The 1993 LucasArts game Zombies Ate My Neighbors features a stage dedicated to Day of the Tentacle. The artists for Day of the Tentacle shared office space with the Zombies Ate My Neighbors development team. The team included the homage after frequently seeing artwork for Day of the Tentacle during the two games' productions. In describing what he considered "the most rewarding moment" of his career, Grossman stated that the game's writing and use of spoken and subtitled dialog assisted a learning-disabled child in learning how to read. Telltale Games CEO Dan Connors commented in 2009 that an episodic game based on Day of the Tentacle was "feasible", but depended on the sales of the Monkey Island games released that year. In 2018, a fan-made sequel, Return of the Tentacle, was released free by a team from Germany. The game imitates the art style of the Remastered edition and features full voice acting. Remasters Special Edition According to Kotaku, a remastered version of Day of the Tentacle was in the works at LucasArts Singapore before the sale of LucasArts to Disney in 2012. Though never officially approved, the game used a pseudo-3D art style and was nearly 80% complete, according to one person close to the project, but was shelved in the days before the closure of LucasArts. Remastered A remastered version of Day of the Tentacle was developed by Schafer and his studio, Double Fine Productions. The remaster was released on March 22, 2016, for PlayStation 4, PlayStation Vita, Microsoft Windows and OS X, with a Linux version released at July 11 together with a mobile port for iOS. The Playstation 4 and Playstation Vita versions are cross-buy and also feature cross-save. An Xbox One port came in October 2020. The remastered game was released as a free PlayStation Plus title for the month of January 2017. Schafer credited both LucasArts and Disney for help in creating the remaster, which follows from a similar remastering of Grim Fandango, as well by Double Fine, in January 2015. Schafer said when they originally were about to secure the rights to Grim Fandango from LucasArts to make the remaster, they did not originally have plans to redo the other LucasArts adventure games, but with the passionate response, they got on the news of the Grim Fandango remaster, they decided to continue these efforts. Schafer described getting the rights to Day of the Tentacle a "miracle" though aided by the fact that many of the executives in the legal rights chain had fond memories of playing these games and helped to secure the rights. 2 Player Productions, which has worked before with Double Fine to document their game development process, also created a mini-documentary for Day of the Tentacle Remastered, which included a visit to the Skywalker Ranch, where LucasArts games were originally developed, where much of the original concept art and digital files for the game and other LucasArts adventure games were archived. Day of the Tentacle Remastered retains its two-dimensional cartoon-style art, redrawn at a higher resolution for modern computers. The high resolution character art was updated by a team led by Yujin Keim with the consultation of Ahern and Chan. Keim's team used many of the original sketches of characters and assets from the two and emulated their style with improvements for modern graphics systems. Matt Hansen worked on recreating the background assets in high resolution. As with the Grim Fandango remaster, the player can switch back and forth between the original graphics and the high-resolution version. The game includes a more streamlined interaction menu, a command wheel akin to the approach used in Broken Age, but the player can opt to switch back to the original interface. The game's soundtrack has been redone within MIDI adapted to work with the iMUSE system. There is an option to listen to commentary from the original creators, including Schafer, Grossman, Chan, McConnell, Ahern, and Bajakian. The remaster contains the fully playable version of the original Maniac Mansion, though no enhancements have been made to that game-within-a-game. Day of the Tentacle Remastered has received positive reviews, with the PC version having an aggregate review score of 87/100 tallied by Metacritic. Reviewers generally praised the game as having not lost its charm since its initial release, but found some aspects of the remastering to be lackluster. Richard Corbett for Eurogamer found the game "every bit as well crafted now as it was in 1993", but found the processes used to provide high-definition graphics from the original 16-bit graphics to making some of the required shortcuts taken in 1993 for graphics, such as background dithering and low animation framerates, more obvious on modern hardware. IGNs Jared Petty also found the remastered to still be enjoyable, and found the improvement on the graphics to be "glorious", but worried that the lack of a hint system, as was added in The Secret of Monkey Island remastered version, would put off new players to the game. Bob Mackey for USgamer found that while past remastered adventure games have highlighted how much has changed in gamers' expectations since the heyday of adventure games in the 1990s, Day of the Tentacle Remastered "rises above these issues to become absolutely timeless". Notes References External links 1993 video games Cultural depictions of Benjamin Franklin Cultural depictions of George Washington Cultural depictions of John Hancock Cultural depictions of Thomas Jefferson DOS games LucasArts games Classic Mac OS games Point-and-click adventure games Science fiction video games SCUMM games ScummVM-supported games Video game sequels Video games about extraterrestrial life Video games about time travel Video games developed in the United States Video games featuring female protagonists Video games scored by Clint Bajakian Video games scored by Michael Land Video games scored by Peter McConnell Video games set in the United States Video games with commentaries Xbox One games
8253
https://en.wikipedia.org/wiki/Dayton%2C%20Ohio
Dayton, Ohio
Dayton () is the sixth-largest city in the U.S. state of Ohio and the county seat of Montgomery County. A small part of the city extends into Greene County. The 2020 U.S. census estimate put the city population at 137,644, while Greater Dayton was estimated to be at 814,049 residents. The Combined Statistical Area (CSA) was 1,086,512. This makes Dayton the fourth-largest metropolitan area in Ohio and 73rd in the United States. Dayton is within Ohio's Miami Valley region, north of the Greater Cincinnati area. Ohio's borders are within of roughly 60 percent of the country's population and manufacturing infrastructure, making the Dayton area a logistical centroid for manufacturers, suppliers, and shippers. Dayton also hosts significant research and development in fields like industrial, aeronautical, and astronautical engineering that have led to many technological innovations. Much of this innovation is due in part to Wright-Patterson Air Force Base and its place in the community. With the decline of heavy manufacturing, Dayton's businesses have diversified into a service economy that includes insurance and legal sectors as well as healthcare and government sectors. Along with defense and aerospace, healthcare accounts for much of the Dayton area's economy. Hospitals in the Greater Dayton area have an estimated combined employment of nearly 32,000 and a yearly economic impact of $6.8 billion. It is estimated that Premier Health Partners, a hospital network, contributes more than $2 billion a year to the region through operating, employment, and capital expenditures. In 2011, Dayton was rated the #3 city in the nation by HealthGrades for excellence in healthcare. Dayton is also noted for its association with aviation; the city is the birthplace of Orville Wright. Other well-known individuals born in the city include poet Paul Laurence Dunbar and entrepreneur John H. Patterson. Dayton is also known for its many patents, inventions, and inventors, most notably the Wright brothers' invention of powered flight. In 2007 Dayton was a part of the top 100 cities in America. In 2008, 2009, and 2010, Site Selection magazine ranked Dayton the #1 mid-sized metropolitan area in the nation for economic development. Also in 2010, Dayton was named one of the best places in the United States for college graduates to find a job. History Dayton was founded on April 1, 1796, by 12 settlers known as the Thompson Party. They traveled in March from Cincinnati up the Great Miami River by pirogue and landed at what is now St. Clair Street, where they found two small camps of Native Americans. Among the Thompson Party was Benjamin Van Cleve, whose memoirs provide insights into the Ohio Valley's history. Two other groups traveling overland arrived several days later. The oldest surviving building is Newcom Tavern, which was used for various purposes, including housing Dayton's first church, which is still in existence. In 1797, Daniel C. Cooper laid out Mad River Road, the first overland connection between Cincinnati and Dayton, opening the "Mad River Country" to settlement. Ohio was admitted into the Union in 1803, and the village of Dayton was incorporated in 1805 and chartered as a city in 1841. The city was named after Jonathan Dayton, a captain in the American Revolutionary War who signed the U.S. Constitution and owned a significant amount of land in the area. In 1827, construction on the Dayton–Cincinnati canal began, which would provide a better way to transport goods from Dayton to Cincinnati and contribute significantly to Dayton's economic growth during the 1800s. Innovation Innovation led to business growth in the region. In 1884, John Henry Patterson acquired James Ritty's National Manufacturing Company along with his cash register patents and formed the National Cash Register Company (NCR). The company manufactured the first mechanical cash registers and played a crucial role in the shaping of Dayton's reputation as an epicenter for manufacturing in the early 1900s. In 1906, Charles F. Kettering, a leading engineer at the company, helped develop the first electric cash register, which propelled NCR into the national spotlight. NCR also helped develop the US Navy Bombe, a code-breaking machine that helped crack the Enigma machine cipher during World War II. Dayton has been the home for many patents and inventions since the 1870s. According to the National Park Service, citing information from the U.S. Patent Office, Dayton had granted more patents per capita than any other U.S. city in 1890 and ranked fifth in the nation as early as 1870. The Wright brothers, inventors of the airplane, and Charles F. Kettering, world-renowned for his numerous inventions, hailed from Dayton. The city was also home to James Ritty's Incorruptible Cashier, the first mechanical cash register, and Arthur E. Morgan's hydraulic jump, a flood prevention mechanism that helped pioneer hydraulic engineering. Paul Laurence Dunbar, an African-American poet and novelist, penned his most famous works in the late 19th century and became an integral part of the city's history. Birthplace of aviation Powered aviation began in Dayton. Orville and Wilbur Wright were the first to construct and demonstrate powered flight. Although the first flight was in Kitty Hawk, North Carolina, their Wright Flyer was built in and returned to Dayton for improvements and further flights at Huffman Field, a cow pasture northeast of Dayton, near the current Wright-Patterson Air Force Base. When the government tried to move development to Langley field in southern Virginia, six Dayton businessmen including Edward A. Deeds, formed the Dayton-Wright Airplane Company in Moraine and established a flying field. Deeds also opened a field to the north in the flood plain of the Great Miami River between the confluences of that river, the Stillwater River, and the Mad River, near downtown Dayton. Later named McCook Field for Alexander McDowell McCook, an American Civil War general, this became the Army Signal Corps' primary aviation research and training location. Wilbur Wright also purchased land near Huffman prairie to continue their research. During World War I, the Army purchased 40 acres adjacent to Huffman Prairie for the Fairfield Aviation General Supply Depot. As airplanes developed more capability, they needed more runway space than McCook could offer, and a new location was sought. The Patterson family formed the Dayton Air Service Committee, Inc which held a campaign that raised $425,000 in two days and purchased northeast of Dayton, including Wilbur Wright Field and the Huffman Prairie Flying Field. Wright Field was "formally dedicated" on October 12, 1927. After World War II, Wright Field and the adjacent Patterson Field, Dayton Army Air Field, and Clinton Army Air Field were merged as the Headquarters, Air Force Technical Base. On January 13, 1948, the facility was renamed Wright-Patterson Air Force Base. Great Dayton Flood A catastrophic flood in March 1913, known as the Great Dayton Flood, led to the creation of the Miami Conservancy District, a series of dams as well as hydraulic pumps installed around Dayton, in 1914. The war effort Like other cities across the country, Dayton was heavily involved in the war effort during World War II. Several locations around the city hosted the Dayton Project, a branch of the larger Manhattan Project, to develop polonium triggers used in early atomic bombs. The war efforts led to a manufacturing boom throughout the city, including high demand for housing and other services. At one point, emergency housing was put into place due to a housing shortage in the region, much of which is still in use today. Alan Turing is widely considered to be the father of theoretical computer science and artificial intelligence. He visited the National Cash Register (NCR) company in Dayton in December 1942. He was able to show that it was not necessary to build 336 Bombes, so the initial order was scaled down to 96 machines to decipher German Enigma-machine-encrypted secret messages during World War II. Post-war Dayton Between the 1940s and the 1970s, the city saw significant growth in suburban areas from population migration. Veterans were returning from military service in large numbers seeking industrial and manufacturing jobs, a part of the local industry that was expanding rapidly. Advancements in architecture also contributed to the suburban boom. New, modernized shopping centers and the Interstate Highway System allowed workers to commute greater distances and families to live further from the downtown area. More than 127,000 homes were built in Montgomery County during the 1950s. During this time, the city was the site of several race riots, including one in 1955 following the murder of Emmett Till, the 1966 Dayton race riot, two in 1967 (following a speech by civil rights activist H. Rap Brown and another following the police killing of an African American man), and one in 1968 as part of the nationwide King assassination riots. Since the 1980s, however, Dayton's population has declined, mainly due to the loss of manufacturing jobs and decentralization of metropolitan areas, as well as the national housing crisis that began in 2008. While much of the state has suffered for similar reasons, the impact on Dayton has been greater than most. Dayton had the third-greatest percentage loss of population in the state since the 1980s, behind Cleveland and Youngstown. Despite this, Dayton has begun diversifying its workforce from manufacturing into other growing sectors such as healthcare and education. Peace accords In 1995, the Dayton Agreement, a peace accord between the parties to the hostilities of the conflict in Bosnia-Herzegovina and the former Yugoslavia, was negotiated at Wright-Patterson Air Force Base, near Fairborn, Ohio, from November 1 to 21. Richard Holbrooke wrote about these events in his memoirs: There was also a real Dayton out there, a charming Ohio city, famous as the birthplace of the Wright brothers. Its citizens energized us from the outset. Unlike the population of, say, New York City, Geneva or Washington, which would scarcely notice another conference, Daytonians were proud to be part of history. Large signs at the commercial airport hailed Dayton as the "temporary center of international peace." The local newspapers and television stations covered the story from every angle, drawing the people deeper into the proceedings. When we ventured into a restaurant or a shopping center downtown, people crowded around, saying that they were praying for us. Warren Christopher was given at least one standing ovation in a restaurant. Families on the airbase placed "candles of peace" in their front windows, and people gathered in peace vigils outside the base. One day they formed a "peace chain," although it was not large enough to surround the sprawling eight-thousand-acre base. Ohio's famous ethnic diversity was on display. 2000s initiatives Downtown expansion that began in the 2000s has helped revitalize the city and encourage growth. Day Air Ballpark, home of the Dayton Dragons, was built in 2000. The highly successful minor league baseball team has been an integral part of Dayton's culture. In 2001, the city's public park system, Five Rivers MetroParks, built RiverScape MetroPark, an outdoor entertainment venue that attracts more than 400,000 visitors each year. A new performance arts theater, the Schuster Center, opened in 2003. A large health network in the region, Premier Health Partners, expanded its Miami Valley Hospital with a 12-story tower addition. In 2010, the Downtown Dayton Partnership, in cooperation with the City of Dayton and community leaders, introduced the Greater Downtown Dayton Plan. It focuses on job creation and retention, infrastructure improvements, housing, recreation, and collaboration. The plan is to be implemented through the year 2020. Nickname Dayton is known as the "Gem City". The nickname's origin is uncertain, but several theories exist. In the early 19th century, a well-known racehorse named Gem hailed from Dayton. In 1845, an article published in the Cincinnati Daily Chronicle by an author known as T stated: In the late 1840s, Major William D. Bickham of the Dayton Journal began a campaign to nickname Dayton the "Gem City." The name was adopted by the city's Board of Trade several years later. Paul Laurence Dunbar referred to the nickname in his poem, "Toast to Dayton", as noted in the following excerpt: She shall ever claim our duty, For she shines—the brightest gem That has ever decked with beauty Dear Ohio's diadem. Dayton also plays a role in a nickname given to the state of Ohio, "Birthplace of Aviation." Dayton is the hometown of the Wright brothers, aviation pioneers who are credited with inventing and building the first practical airplane in history. After their first manned flights in Kitty Hawk, North Carolina, which they had chosen due to its ideal weather and climate conditions, the Wrights returned to Dayton and continued testing at nearby Huffman Prairie. Additionally, Dayton is colloquially referred to as "Little Detroit". This nickname comes from Dayton's prominence as a Midwestern manufacturing center. Geography According to the United States Census Bureau, the city has a total area of , of which is land and is water. Climate Dayton's climate features warm, muggy summers and cold, dry winters, and is classified as a humid continental climate (Köppen Dfa). Unless otherwise noted, all normal figures quoted within the text below are from the official climatology station, Dayton International Airport, at an elevation of about to the north of downtown Dayton, which lies within the valley of the Miami River; thus temperatures there are typically cooler than in downtown. At the airport, monthly mean temperatures range from in January to in July. The highest temperature ever recorded in Dayton was on July 22, 1901, and the coldest was on February 13 during the Great Blizzard of 1899. On average, there are 14 days of + highs and 4.5 nights of sub- lows annually. Snow is moderate, with a normal seasonal accumulation of , usually occurring from November to March, occasionally April, and rarely October. Precipitation averages annually, with total rainfall peaking in May. Dayton is subject to severe weather typical of the Midwestern United States. Tornadoes are possible from the spring to the fall. Floods, blizzards, and severe thunderstorms can also occur. On Memorial Day of 2019, Dayton suffered extensive property damage and one death during a tornado outbreak, in which a total of 15 tornadoes touched down in the Dayton area. Although some of the tornadoes were only EF0 and remained on the ground for less than a mile, one was an EF4 measuring a half-mile-wide (805 meters), which tore through the communities of Brookville, Trotwood, Dayton, and Riverside. Several streets were closed, including portions of I-75 and North Dixie Street. 64,000 residents lost power and much of the region's water supply was cut off. Ecology The Dayton Audubon Society is the National Audubon Society's local chapter. The Dayton chapter manages local activities contributing to the annual, hemisphere-wide Christmas Bird Count. The Chapter began participation in the National Count in 1924. The local Count was initially coordinated by Ben Blincoe, who was succeeded by Jim Hill in 1970. In the mid-1960s, the freezing of Lake Erie and associated marshlands led species of waterfowl to appear in the Dayton-area, where surface waters remained unfrozen. Nine varieties of birds have been observed every year in the Dayton area: downy woodpecker, Carolina chickadee, tufted titmouse, brown creeper, cardinal, junco, tree sparrow, song sparrow and crow. Demographics Note: the following demographic information applies only to the city of Dayton proper. For other Dayton-area communities, see their respective articles. Dayton's population declined significantly from a peak of 262,332 residents in 1960 to only 141,759 in 2010. This was in part due to the slowdown of the region's manufacturing and the growth of Dayton's affluent suburbs including Oakwood, Englewood, Beavercreek, Springboro, Miamisburg, Kettering, and Centerville. The city's most populous ethnic group, white, declined from 78.1% in 1960 to 51.7% by 2010. Recent census estimates show a population decline since 2010. As of the 2000 census, the median income for a household in the city was $27,523, and the median income for a family was $34,978. Males had a median income of $30,816 versus $24,937 for females. The per capita income for the city was $34,724. About 18.2% of families and 23.0% of the population were below the poverty line, including 32.0% of those under age 18 and 15.3% of those age 65 or over. 2010 census As of the 2010 census, there were 141,759 people, 58,404 households, and 31,064 families residing in the city. The population density was . There were 74,065 housing units at an average density of . The racial makeup of the city was 51.7% White, 42.9% African American, 0.3% Native American, 0.9% Asian, 1.3% from other races, and 2.9% from two or more races. Hispanic or Latino of any race were 3.0% of the population. There were 58,404 households, of which 28.3% had children under the age of 18 living with them, 25.9% were married couples living together, 21.4% had a female householder with no husband present, 5.9% had a male householder with no wife present, and 46.8% were non-families. 38.8% of all households were made up of individuals, and 11.2% had someone living alone who was 65 years of age or older. The average household size was 2.26, and the average family size was 3.03. The median age in the city was 34.4 years. 22.9% of residents were under the age of 18; 14.2% were between the ages of 18 and 24; 25.3% were from 25 to 44; 25.8% were from 45 to 64, and 11.8% were 65 years of age or older. The gender makeup of the city was 48.7% male and 51.3% female. 2013 census population estimates The 2013 census population estimate showed a slight Dayton population increase for the first time in five decades. However, the 2014 population estimate indicated a decrease of individuals from 2013's estimate. Economy Dayton's economy is relatively diversified and vital to the overall economy of the state of Ohio. In 2008 and 2009, Site Selection magazine ranked Dayton the #1 medium-sized metropolitan area in the U.S. for economic development. Dayton is also among the top 100 metropolitan areas in both exports and export-related jobs, ranked 16 and 14 respectively by the Brookings Institution. The 2010 report placed the value of exports at $4.7 billion and the number of export-related jobs at 44,133. The Dayton Metropolitan Statistical Area ranks 4th in Ohio's Gross Domestic Product with a 2008 industry total of $33.78 billion. Additionally, Dayton ranks third among 11 major metropolitan areas in Ohio for exports to foreign countries. The Dayton Development Coalition is attempting to leverage the region's large water capacity, estimated to be 1.5 trillion gallons of renewable water aquifers, to attract new businesses. Moody's Investment Services revised Dayton's bond rating from A1 to the stronger rating of Aa2 as part of its global recalibration process. Standard & Poor's upgraded Dayton's rating from A+ to AA- in the summer of 2009. Bloomberg Businessweek ranked Dayton in 2010 as one of the best places in the U.S. for college graduates looking for a job. Companies such as Reynolds and Reynolds, CareSource, DP&L (soon AES inc), LexisNexis, Kettering Health Network, Premier Health Partners, and Standard Register have their headquarters in Dayton. It is also the former home of the Speedwell Motor Car Company, MeadWestvaco (formerly known as the Mead Paper Company), and NCR. NCR was headquartered in Dayton for over 125 years and was a major innovator in computer technology. Research, development, aerospace and aviation The Dayton region gave birth to aviation and is known for its high concentration of aerospace and aviation technology. In 2009, Governor Ted Strickland designated Dayton as Ohio's aerospace innovation hub, the state's first such technology hub. Two major United States research and development organizations have leveraged Dayton's historical leadership in aviation and maintain their headquarters in the area: The National Air and Space Intelligence Center (NASIC) and the Air Force Research Laboratory (AFRL). Both have their headquarters at Wright-Patterson Air Force Base. Several research organizations support NASIC, AFRL, and the Dayton community. The Advanced Technical Intelligence Center is a confederation of government, academic, and industry partners. The University of Dayton Research Institute (UDRI) is led by the University of Dayton. The Cognitive Technologies Division (CTD) of Applied Research Associates, Inc., which carries out human-centered research and design, is headquartered in the Dayton suburb of Fairborn. The city of Dayton has started Tech Town, a development project to attract technology-based firms and revitalize the downtown area. Tech Town is home to the world's first RFID business incubator. The University of Dayton–led Institute for Development & Commercialization of Sensor Technologies (IDCAST) at TechTown is a center for remote sensing and sensing technology. It is one of Dayton's technology business incubators housed in The Entrepreneurs Center building. Healthcare The Kettering Health Network and Premier Health Partners have a major role on the Dayton area's economy. Hospitals in the Greater Dayton area have an estimated combined employment of nearly 32,000 and a yearly economic impact of $6.8 billion. In addition, several Dayton area hospitals consistently earn top national ranking and recognition including the U.S. News & World Reports list of "America's Best Hospitals" as well as many of HealthGrades top ratings. The most notable hospitals are Miami Valley Hospital and Kettering Medical Center. The Dayton region has several key institutes and centers for health care. The Center for Tissue Regeneration and Engineering at Dayton focuses on the science and development of human tissue regeneration. The National Center for Medical Readiness (NCMR) is also in the Dayton area. The center includes Calamityville, which is a disaster training facility. Over five years, Calamityville is estimated to have a regional economic impact of $374 million. Also, the Neurological Institute at Miami Valley Hospital is an institute focused on the diagnosis, treatment, and research of neurological disorders. Top employers According to the city's 2019 Comprehensive Annual Financial Report, the top employers in the city proper are: Government The Dayton City Commission is composed of the mayor and four city commissioners. Each city commission member is elected at-large on a non-partisan basis for four-year, overlapping terms. All policy items are decided by the city commission, which is empowered by the City Charter to pass ordinances and resolutions, adopt regulations, and appoint the city manager. The city manager is responsible for budgeting and implementing policies and initiatives. Dayton was the first large American city to adopt the city manager form of municipal government, in 1913. Cityscape Architecture Unlike many Midwestern cities its age, Dayton has very broad and straight downtown streets (generally two or three full lanes in each direction) that improved access to the downtown even after the automobile became popular. The main reason for the broad streets was that Dayton was a marketing and shipping center from its beginning; streets were broad to enable wagons drawn by teams of three to four pairs of oxen to turn around. Also, some of today's streets were once barge canals flanked by draw-paths. A courthouse building was built in downtown Dayton in 1888 to supplement Dayton's original Neoclassical courthouse, which still stands. This second, "new" courthouse has since been replaced with new facilities as well as a park. The Old Court House has been a favored political campaign stop. On September 17, 1859, Abraham Lincoln delivered an address on its steps. Eight other presidents have visited the courthouse, either as presidents or during presidential campaigns: Andrew Johnson, James Garfield, John F. Kennedy, Lyndon B. Johnson, Richard Nixon, Gerald Ford, Ronald Reagan, and Bill Clinton. The Dayton Arcade, which opened on March 3, 1904, was built in the hopes of replacing open-air markets throughout the city. Throughout the decades, the Arcade has gone through many transformations but has retained its charm. Some of its main features include a Flemish facade at the Third Street entrance, a glass dome above the Arcade rotunda, and a chateau roof line above the Third Street facade. The Dayton Arcade is currently under renovations with no official completion date set. In 2009, the CareSource Management Group finished construction of a $55 million corporate headquarters in downtown Dayton. The , 10-story building was downtown's first new office tower in more than a decade. Dayton's two tallest buildings are the Kettering Tower at and the KeyBank Tower at . Kettering Tower was originally Winters Tower, the headquarters of Winters Bank. The building was renamed after Virginia Kettering when Winters was merged into Bank One. KeyBank Tower was known as the MeadWestvaco Tower before KeyBank gained naming rights to the building in 2008. Ted Rall said in 2015 that over the last five decades Dayton has been demolishing some of its architecturally significant buildings to reduce the city's rental vacancy rate and thus increase the occupancy rate. Neighborhoods Dayton's ten historic neighborhoods—Oregon District, Wright Dunbar, Dayton View, Grafton Hill, McPherson Town, Webster Station, Huffman, Kenilworth, St. Anne's Hill, and South Park—feature mostly single-family houses and mansions in the Neoclassical, Jacobethan, Tudor Revival, English Gothic, Chateauesque, Craftsman, Queen Anne, Georgian Revival, Colonial Revival, Renaissance Revival Architecture, Shingle Style Architecture, Prairie, Mission Revival, Eastlake/Italianate, American Foursquare, and Federal styles. Downtown Dayton is also a large area that encompasses several neighborhoods itself and has seen a recent uplift and revival. Suburbs Dayton's suburbs with a population of 10,000 or more include Beavercreek, Centerville, Clayton, Englewood, Fairborn, Harrison Township, Huber Heights, Kettering, Miami Township, Miamisburg, Oakwood, Riverside, Springboro, Trotwood, Vandalia, Washington Township, West Carrollton, and Xenia. In the federal government's National Urban Policy and New Community Development Act of 1970, funding was provided for thirteen "new towns" or planned cities throughout the country. One location was set to become a suburb of Dayton and was known variously as Brookwood or Newfields. The goal was to have an entirely new suburb that would eventually house about 35,000 residents. The new town was to be located between Trotwood and Brookville, and modeled on the ideas of Ian McHarg. The project was abandoned in 1978 and most of the land became Sycamore State Park. Recreation Dayton was named National Geographic's outdoor adventure capital of the Midwest in 2019 due in large part to the metropolitan area's revitalized Five Rivers MetroPark, extensive bicycle and jogging trail system, urban green spaces, lakes and camping areas. Bicycling In cooperation with the Miami Conservancy District, Five Rivers MetroParks hosts 340 miles of paved trails, the largest network of paved off-street trails in the United States. The regional trail system represents over 35% of the 900 miles in Ohio's off-street trail network. In 2010, the city of Troy was named "bike friendly" by the League of American Bicyclists, which gave the city the organization's bronze designation. The honorable mention made Dayton one of two cities in Ohio to receive the award, the other being Columbus, and one of 15 cities nationwide. Culture Fine arts The Dayton Region ranked within the top 10% in the nation in arts and culture. In a 2012 readers' poll by American Style magazine, Dayton ranked #2 in the country among mid-size cities as an arts destination, ranking higher than larger cities such as Atlanta, St. Louis, and Cincinnati. Dayton is the home of the Dayton Art Institute. The Benjamin and Marian Schuster Performing Arts Center in downtown Dayton is a world-class performing arts center and the home venue of the Dayton Philharmonic Orchestra, Dayton Opera, and the Dayton Ballet. In addition to philharmonic and opera performances, the Schuster Center hosts concerts, lectures, and traveling Broadway shows, and is a popular spot for weddings and other events. The historic Victoria Theatre in downtown Dayton hosts concerts, traveling Broadway shows, ballet, a summertime classic film series, and more. The Loft Theatre, also downtown, is the home of the Human Race Theatre Company. The Dayton Playhouse, in West Dayton, is the site of numerous plays and theatrical productions. Between 1957 and 1995, the Kenley Players presented live theater productions in Dayton. In 2013, John Kenley was inducted into the Dayton Theatre Hall of Fame. Dayton is the home to several ballet companies including: The Dayton Ballet, one of the oldest professional dance companies in the United States. The Dayton Ballet runs the Dayton Ballet School, the oldest dance school in Dayton and one of the oldest in the country. It is the only ballet school in the Miami Valley associated with a professional dance company. The Dayton Contemporary Dance Company (established in 1968), which hosts the largest repertory of African-American-based contemporary dance in the world. The company travels nationally and internationally and has been recognized by critics worldwide. Front Street, the largest artists' collective in Dayton, is housed in three industrial buildings on East Second Street. Food The city's fine dining restaurants include The Pine Club, a nationally known steakhouse. Dayton is home to a variety of pizza chains that have become woven into local culture, the most notable of which are Cassano's and Marion's Piazza. Notable Dayton-based restaurant chains include Hot Head Burritos. In addition to restaurants, the city is also home to Esther Price Candies, a candy and chocolate company, and Mike-sells, the oldest potato chip company in the United States. The city began developing a reputation for its number of breweries and craft beer venues by the late 2010s. Religion Many major religions are represented in Dayton. Christianity is represented in Dayton by dozens of denominations and their respective churches. Notable Dayton churches include the First Lutheran Church, Sacred Heart Church, and Ginghamsburg Church. Dayton's Muslim community is largely represented by the Islamic Society of Greater Dayton (ISGD), a Muslim community that includes a mosque on Josie Street. Dayton is also home to the United Theological Seminary, one of 13 seminaries affiliated with the United Methodist Church. Judaism is represented by Temple Israel. Hinduism is represented by the Hindu Temple of Dayton. Old North Dayton also has a number of Catholic churches built by immigrants from Lithuania, Poland, Hungary, and Germany. Tourism Tourism also accounts for one out of every 14 private sector jobs in the county. Tourism in the Dayton region is led by the National Museum of the United States Air Force at Wright-Patterson Air Force Base, the largest and oldest military aviation museum in the world. The museum draws over 1.3 million visitors per year and is one of the most-visited tourist attractions in Ohio. The museum houses the National Aviation Hall of Fame. Other museums also play significant roles in the tourism and economy of the Dayton area. The Dayton Art Institute, a museum of fine arts, owns collections containing more than 20,000 objects spanning 5,000 years of art and archaeological history. The Dayton Art Institute was rated one of the top 10 best art museums in the United States for children. The Boonshoft Museum of Discovery is a children's museum of science with numerous exhibits, one of which includes an indoor zoo with nearly 100 different animals. There are also some notable historical museums in the region. The Dayton Aviation Heritage National Historical Park, operated by the National Park Service, commemorates the lives and achievements of Dayton natives Orville and Wilbur Wright and Paul Laurence Dunbar. The Wright brothers' famous Wright Flyer III aircraft is housed in a museum at Carillon Historical Park. Dayton is also home to America's Packard Museum, which contains many restored historical Packard vehicles. SunWatch Indian Village/Archaeological Park, a partially reconstructed 12th-century prehistoric American Indian village, is on the south end of Dayton; it is organized around a central plaza dominated by wood posts forming an astronomical calendar. The park includes a museum where visitors can learn about the Indian history of the Miami Valley. Music & Entertainment The Vectren Dayton Air Show is an annual air show that takes place at the Dayton International Airport. The Vectren Dayton Airshow is one of the largest air shows in the United States. The Dayton area is served by Five Rivers MetroParks, encompassing over 23 facilities for year-round recreation, education, and conservation. In cooperation with the Miami Conservancy District, the MetroParks maintains over of paved, multi-use scenic trails that connect Montgomery County with Greene, Miami, Warren, and Butler counties. Dayton was home to a thriving funk music scene from the 1970s to the early 1980s, that included bands such as Ohio Players, Roger Troutman & Zapp, Lakeside, Sun, Dayton, Heatwave, and Slave. Dayton was also the birthplace to several influential indie and punk bands such as The Breeders, Guided by Voices, and Brainiac. From 1996 to 1998, Dayton hosted the National Folk Festival. Since then, the annual Cityfolk Festival has continued to bring folk, ethnic, and world music and arts to Dayton. The Five Rivers MetroParks also owns and operates the PNC Second Street Market near downtown Dayton. The Dayton area hosts several arenas and venues. South of Dayton in Kettering is the Fraze Pavilion, whose notable performances have included the Backstreet Boys, Boston, and Steve Miller Band. South of downtown, on the banks of the Great Miami River, is the University of Dayton Arena, home venue for the University of Dayton Flyers basketball teams and the location of various other events and concerts. It also hosts the Winter Guard International championships, at which hundreds of percussion and color guard ensembles from around the world compete. In addition, the Dayton Amateur Radio Association hosts the annual Dayton Hamvention, North America's largest hamfest, at the Greene County Fairgrounds in nearby Xenia. The Nutter Center, which is just east of Dayton in the suburb of Fairborn, is the home arena for athletics of Wright State University and the former Dayton Bombers hockey team. This venue is used for many concerts, community events, and various national traveling shows and performances. The Oregon District is a historic residential and commercial district in southeast downtown Dayton. The district is populated with art galleries, specialty shops, pubs, nightclubs, and coffee houses. The city of Dayton is also host to yearly festivals, such as the Dayton Celtic Festival, the Dayton Blues Festival, Dayton Music Fest, Urban Nights, Women in Jazz, the African American and Cultural Festival, and the Dayton Reggae Fest. Sports The Dayton area is home to several minor league and semi pro teams, as well as NCAA Division I sports programs. Baseball The Dayton Dragons professional baseball team is a Class A minor league affiliate for the Cincinnati Reds. The Dayton Dragons are the first (and only) team in minor league baseball history to sell out an entire season before it began and was voted as one of the top 10 hottest tickets to get in all of professional sports by Sports Illustrated. The Dayton Dragons 815 consecutive sellouts surpassed the NBA's Portland Trail Blazers for the longest sellout streak across all professional sports in the U.S. Collegiate The University of Dayton and Wright State University both host NCAA basketball. The University of Dayton Arena has hosted more games in the NCAA men's basketball tournament over its history than any other venue. UD Arena is also the site of the First Round games of the NCAA Tournament. In 2012, eight teams competed for the final four spots in the NCAA basketball tournament. Wright State University's NCAA men's basketball is the Wright State Raiders and the University of Dayton's NCAA men's basketball team is the Dayton Flyers. Hockey The Dayton Gems were a minor league ice hockey team in the International Hockey League from 1964 to 1977, 1979 to 1980, and most recently 2009 to 2012. The Dayton Bombers were an ECHL ice hockey team from 1991 to 2009. They most recently played the North Division of the ECHL's American Conference. In June 2009, it was announced the Bombers would turn in their membership back to the league. Despite the folding of the Bombers, hockey remained in Dayton as the Dayton Gems of the International Hockey League were formed in the fall of 2009 at Hara Arena. The Gems folded after the 2011–12 season. Shortly after the Gems folded, it was announced a new team, the Dayton Demonz, would begin play in 2012 in the Federal Hockey League (FHL). The Demonz folded in 2015 and were immediately replaced by the Dayton Demolition, also in the FHL. However, the Demolition would cease operations after only one season when Hara Arena decided to close due to financial difficulties. Football Dayton hosted the first American Professional Football Association game (precursor to the NFL). The game was played at Triangle Park between the Dayton Triangles and the Columbus Panhandles on October 3, 1920, and is considered one of the first professional football games ever played. Football teams in the Dayton area include the Dayton Flyers and the Dayton Sharks. Golf The Dayton region is also known for the many golf courses and clubs that it hosts. The Miami Valley Golf Club, Moraine Country Club, NCR Country Club, and the Pipestone Golf Course are some of the more notable courses. Also, several PGA Championships have been held at area golf courses. The Miami Valley Golf Club hosted the 1957 PGA Championship, the Moraine Country Club hosted the 1945 PGA Championship, and the NCR Country club hosted the 1969 PGA Championship. Additionally, NCR CC hosted the 1986 U.S. Women's Open and the 2005 U.S. Senior Open. Other notable courses include the Yankee Trace Golf Club, the Beavercreek Golf Club, Dayton Meadowbrook Country Club, Sycamore Creek Country Club, Heatherwoode Golf Club, Community Golf Course, and Kitty Hawk Golf Course. Rugby The city of Dayton is the home to the Dayton Area Rugby Club which hosts their home games at the Dayton Rugby Grounds. As of 2018, the club fields two men's and one women's side for Rugby Union and several Rugby Sevens sides. The club also hosts the annual Gem City 7's tournament. Media Newspapers Dayton is served in print by The Dayton Daily News, the city's sole remaining daily newspaper. The Dayton Daily News is owned by Cox Enterprises. The Dayton region's main business newspaper is the Dayton Business Journal. The Dayton City Paper, a community paper focused on music, art, and independent thought ceased operation in 2018. The Dayton Weekly News has been published since 1993, providing news and information to Dayton's African-American community. Magazines There are numerous magazines produced in and for the Dayton region. The Dayton Magazine provides insight into arts, food, and events. Focus on Business is published by the Chamber of Commerce to provide awareness of companies and initiatives affecting the regional economy Television Nielsen Media Research ranked the 11-county Dayton television market as the No. 62 market in the United States. The market is served by stations affiliated with major American networks including: WDTN, channel 2 – NBC, operated by Nexstar Media Group; WHIO-TV, channel 7 – CBS, operated by Cox Media Group; WPTD, channel 16 – PBS, operated by ThinkTV, which also operates WPTO, assigned to Oxford; WKEF, channel 22 – ABC/Fox, operated by Sinclair Broadcasting; WBDT, channel 26 – The CW, operated by Vaughan Media (a shell corporation of Nexstar), assigned to Springfield; WKOI-TV, channel 43 – Ion Television, assigned to Richmond, Indiana; and WRGT-TV, channel 45 – My Network TV, operated under a local marketing agreement by Sinclair Broadcasting. The nationally syndicated morning talk show The Daily Buzz originated from WBDT, the former ACME Communications property in Miamisburg, before moving to its current home in Florida. Radio Dayton is also served by 42 AM and FM radio stations directly, and numerous other stations are heard from elsewhere in southwest Ohio, which serve outlying suburbs and adjoining counties. Transportation Public transit The Greater Dayton Regional Transit Authority (RTA) operates public bus routes in the Dayton metro area. In addition to routes covered by traditional diesel-powered buses, RTA has several electric trolley bus routes. The Dayton trolleybus system is the second longest-running of the five remaining trolleybus systems in the U.S., having entered service in 1933. It is the present manifestation of an electric transit service that has operated continuously in Dayton since 1888. Dayton operates a Greyhound Station which provides inter-city bus transportation to and from Dayton. The hub is in the Greater Dayton Regional Transit Authority North-West hub in Trotwood. Airports Dayton International Airport lies in a northern exclave of the city and offers service to 21 markets through 10 airlines. In 2008, it served 2.9 million passengers. The Dayton International Airport is also a significant regional air freight hub hosting FedEx Express, UPS Airlines, United States Postal Service, and major commercial freight carriers. The Dayton area also has several regional airports. The Dayton–Wright Brothers Airport is a general aviation airport owned by the City of Dayton south of the central business district of Dayton on Springboro Pike in Miami Township. It serves as the reliever airport for Dayton International Airport. The airport primarily serves corporate and personal aircraft users. The Dahio Trotwood Airport, also known as Dayton-New Lebanon Airport, is a privately owned, public-use airport west of the central business district of Dayton. The Moraine Airpark is a privately owned, public-use airport southwest of the city of Dayton. Major highways The Dayton region is primarily served by three interstates: Interstate 75 runs north to south through the city of Dayton and many of Dayton's north and south suburbs, including Kettering and Centerville south of Dayton and Vandalia, Tipp City, and Troy north of Dayton. Interstate 70 is a major east–west interstate that runs through many of Dayton's east and west suburbs, including Huber Heights, Butler Township, Englewood, and Brookville, and intersects with I-75 in Vandalia, Ohio, just north of the city. This intersection of I-70 and I-75 is also known as "Freedom Veterans Crossroads", which was officially named by the U.S. Department of Transportation in 2004. I-70 is the major route to the airport. Interstate 675 is a partial interstate ring on the southeastern and eastern suburbs of Dayton. It runs northeast to south and connects to I-70 to the northeast and I-75 to the south. Other major routes for the region include: US 35 is a major limited access east–west highway that bisects the city. It is most widely used between Drexel and Xenia. Route 40 is a major east–west highway that runs parallel to (and 2 miles north of) I-70 State Route 4 is a freeway that is most heavily traveled between I-75 and I-70. State Route 444 is north–south state highway. Its southern terminus is at its interchange with Route 4, and its northern terminus is at Interstate 675. This limited-access road serves Dayton and Fairborn and is a significant route to access points serving Wright-Patterson Air Force Base. From 2010 through 2017, the Ohio Department of Transportation (ODOT) performed a $533 million construction project to modify, reconstruct and widen I-75 through downtown Dayton, from Edwin C Moses Blvd. to Stanley Avenue. Rail Dayton hosts several inter-modal freight railroad terminals. Two Class I railroads, CSX and Norfolk Southern Railway, operate switching yards in the city. Formerly the Baltimore & Ohio Railroad, New York Central Railroad and the Pennsylvania Railroad, and afterward, Amtrak made long-distance passenger train stops at Dayton Union Station on S. Sixth Street. The last train leaving there was the National Limited in October 1979. Education Public schools Dayton Public Schools operates 34 schools that serve 16,855 students, including: Belmont High Meadowdale High Paul Laurence Dunbar High Ponitz Career Technology Center Stivers School for the Arts Thurgood Marshall High Private schools The city of Dayton has more than 35 private schools within the city, including: Archbishop Alter High School Carroll High School Chaminade Julienne Catholic High School Dayton Christian School Dominion Academy of Dayton The Miami Valley School Spring Valley Academy Charter schools Dayton has 33 charter schools. Three of the top five charter schools named in 2011 are K–8 schools managed by National Heritage Academies. Notable charter schools include: Dayton Early College Academy Emerson Academy North Dayton School of Discovery Pathway School of Discovery Richard Allen Schools Colleges and universities The Dayton area was ranked tenth for higher education among metropolitan areas in the United States by Forbes in 2009. The city is home to two major universities. The University of Dayton is a private, Catholic institution founded in 1850 by the Marianist order. It has the only American Bar Association (ABA)-approved law school in the Dayton area. The University of Dayton is Ohio's largest private university and is also home to the University of Dayton Research Institute, which ranks third in the nation for sponsored materials research, and the Center for Tissue Regeneration and Engineering at Dayton, which focuses on human tissue regeneration. The public Wright State University became a state university in 1967. Wright State University established the National Center for Medical Readiness, a national training program for disaster preparedness and relief. Wright State's Boonshoft School of Medicine is the Dayton area's only medical school and is a leader in biomedical research. Dayton is also home to Sinclair Community College, the largest community college at a single location in Ohio and one of the nation's largest community colleges. Sinclair is acclaimed as one of the country's best community colleges. Sinclair was founded as the YMCA college in 1887. Other schools just outside Dayton that shape the educational landscape are Antioch College and Antioch University, both in Yellow Springs, Central State University in Wilberforce, Kettering College of Medical Arts and School of Advertising Art in Kettering, DeVry University in Beavercreek, Cedarville University, Clark State Community College and Wittenberg University in Springfield. The Air Force Institute of Technology, which was founded in 1919 and serves as a graduate school for the United States Air Force, is at the nearby Wright-Patterson Air Force Base. Institutions Boonshoft School of Medicine Dayton Art Institute Ohio Institute of Photography and Technology School of Advertising Art Wright State University Crime Dayton's crime declined between 2003 and 2008 in key categories according to FBI Uniform Crime Reports and Dayton Police Department data. In 2009, crime continued to fall in the city of Dayton. Crime in the categories of forcible rape, aggravated assault, property crime, motor vehicle theft, robbery, burglary, theft and arson all showed declines for 2009. Overall, crime in Dayton dropped 40% over the previous year. The Dayton Police Department reported a total of 39 murders in 2016, which marked a 39.3% increase in homicides from 2015. John Dillinger, a bank robber during the early 1930s, was captured and arrested by Dayton city police while visiting his girlfriend at a high-class boarding house in downtown Dayton. On August 4, 2019, a mass shooting took place in Dayton. Ten people were killed, including the perpetrator; and twenty-seven were injured. Sister cities Dayton's sister cities are: Sarajevo, Bosnia and Herzegovina Augsburg, Germany Holon, Israel Monrovia, Liberia Ōiso, Japan Rushmoor, England Notable people See also List of mayors of Dayton, Ohio List of people from Dayton, Ohio List of U.S. cities with large Black populations National Aviation Hall of Fame Politics of Dayton, Ohio USS Dayton, 2 ships Explanatory notes References Further reading Conover, Charlotte Reeve. Dayton, Ohio : an intimate history (1995) online Drury, Augustus Waldo. History of the City of Dayton and Montgomery County, Ohio (S. J. Clarke Publishing Company, 1909). online Funk, Nellis R. A Pictorial History of the Great Dayton Flood, March 25, 26, 27, 1913 (1913) online; a primary source Millsap, Adam. "How the Gem city lost its luster and how it can get it back: A case study of Dayton, Ohio." Mercatus Research Paper (2017). online Pocock, Emil. "Popular Roots of Jacksonian Democracy: The Case of Dayton, Ohio, 1815-1830." Journal of the Early Republic 9.4 (1989): 489–515. online Sealander, Judith. Grand Plans: Business Progressivism and Social Change in Ohio's Miami Valley, 1890-1929 (1988) on Dayton and surrounding region. Sharts, Joseph W. Biography Of Dayton - An Economic Interpretation of Local History (1922) online Walker, John T. "Socialism in Dayton, Ohio, 1912 to 1925: Its membership, organization, and demise." Labor History 26.3 (1985): 384–404. Watras, Joseph. "The Racial Desegregation of Dayton, Ohio, Public Schools, 1966–2008." Ohio History 117.1 (2010): 93–107. online External links City website Dayton Area Chamber of Commerce Greater Dayton CVB 1796 establishments in the Northwest Territory Cities in Greene County, Ohio Cities in Montgomery County, Ohio Cities in Ohio County seats in Ohio Populated places established in 1796 Wright brothers World War II Heritage Cities
8419
https://en.wikipedia.org/wiki/Dartmouth%2C%20Devon
Dartmouth, Devon
Dartmouth () is a town and civil parish in the English county of Devon. It is a tourist destination set on the western bank of the estuary of the River Dart, which is a long narrow tidal ria that runs inland as far as Totnes. It lies within the South Devon Area of Outstanding Natural Beauty and South Hams district, and had a population of 5,512 in 2001, reducing to 5,064 at the 2011 census. There are two electoral wards in the Dartmouth area (Townstal & Kingswear). Their combined population at the above census was 6,822. History In 1086, the Domesday Book listed Dunestal as the only settlement in the area which now makes up the parish of Dartmouth. It was held by Walter of Douai. It paid tax on half a hide, and had two plough teams, two slaves, five villagers and four smallholders. There were six cattle, 40 sheep and 15 goats. At this time Townstal (as the name became) was apparently a purely agricultural settlement, centred around the church. Walter of Douai rebelled against William II, and his lands were confiscated and added to the honour of Marshwood (Dorset), which sublet Townstal and Dartmouth to the FitzStephens. It was probably during the early part of their proprietorship that Dartmouth began to grow as a port, as it was of strategic importance as a deep-water port for sailing vessels. The port was used as the sailing point for the Crusades of 1147 and 1190, and Warfleet Creek, close to Dartmouth Castle is supposed by some to be named for the vast fleets which assembled there. Dartmouth was a home of the Royal Navy from the reign of Edward III and was twice surprised and sacked during the Hundred Years' War, after which the mouth of the estuary was closed every night with a great chain. The narrow mouth of the Dart is protected by two fortified castles, Dartmouth Castle and Kingswear Castle. Originally Dartmouth's only wharf was Bayard's Cove, a relatively small area protected by a fort at the southern end of the town. In 1373 Geoffrey Chaucer visited and among the pilgrims in his Canterbury Tales, Notwithstanding Dartmouth's connections with the crown and respectable society, it was a major base for privateering in medieval times. John Hawley or Hauley, a licensed privateer and sometime mayor of Dartmouth is reputed to be a model for Chaucer's "schipman". The earliest street in Dartmouth to be recorded by name (in the 13th century) is Smith Street. Several of the houses on the street are originally late 16th century or early 17th century and probably rebuilt on the site of earlier medieval dwellings. The street name undoubtedly derives from the smiths and shipwrights who built and repaired ships here when the tidal waters reached as far as this point. Smith Street was also the site of the town pillory in medieval times. The first church in the parish was St Clement's, Townstal, which may have existed in some form before the 1190s. It was granted by the FitzStephens to Torre Abbey in about 1198, the Abbey having been founded in 1196, and the present stone-built church was probably started shortly after this. Manorial transactions are first recorded in 1220, when the manor house was at Norton, about half a mile west of Townstal. Names of occupations also started to appear, including taverner, tailor, coggar, korker, goldsmith, glover, skinner and baker. The "Fosse", now Foss Street, a dam across the creek known later as The Mill Pool, was first mentioned in 1243. The flow of water out of the pool through the Mill Gullet powered a tidal mill. The dam was used as an unofficial footpath linking Clifton, to the south, with Hardness, to the north. Before this it was necessary to go westwards to the head of the creek at Ford to travel between the two settlements. The lord of the manor was given the rights to hold a weekly market and an annual fair in 1231. In 1281, a legal case proved that the Lord of Totnes had the right to charge tolls on ships using the river, and this right was bought by Nicholas of Tewkesbury in 1306, who conveyed the town, river and port to the king in 1327, so making Dartmouth a Royal Borough. The king gave the river to the Duchy of Cornwall in 1333, who still own the "fundus" or bed of the river. In 1335 Edward III granted Dartmouth to Joan of Carew, whose husband was Lord of Stoke Fleming, and almost immediately she obediently passed the lordship to Guy de Bryan, one of the king's leading ministers. In 1341, the town was granted a Royal Charter, which allowed for the election of a mayor. The borough was required to provide two ships for forty days per year. After 1390, no more is heard of lordship rights, and the borough became effectively independent of any lord. St Saviour's Church was constructed in 1335 and consecrated in 1372. It contains a pre-Reformation oak rood screen built in 1480 and several monuments including the tomb of John Hawley (died 1408) and his two wives, covered with a large brass plate effigy of all three. A large medieval ironwork door is decorated with two leopards of the Plantagenets and is possibly the original portal. Although it is dated "1631", this is thought to be the date of a subsequent refurbishment coincidental with major renovations of the church in the 17th century. The gallery of the church is decorated with the heraldic crests of prominent local families and is reputed to be constructed of timbers from ships captured during the defeat of the Spanish Armada, although this has not been categorically substantiated. An engraving of a painting by Thomas Allom of the interior of the church, showing the rood screen, provided the inspiration for Letitia Elizabeth Landon's poetical illustration Dartmouth Church in Fisher's Drawing Room scrap Book, 1833. In medieval times, land access from the Totnes direction passed the manor at Norton and the parish church at Townstal before falling steeply along what are now Church Road, Mount Boone and Ridge Hill to the river at Hardness. There were steeper routes via Townstal Hill and Clarence Street and also via Brown's Hill. These were all too steep for vehicles, so the only land access was by packhorse. In 1671 there is the first mention of the building of the "New Ground". A previously existing sandbank was built up using ships' ballast, and a quay wall was built around it to provide more mooring space. The area proved too unstable to be built on, and is now the Royal Avenue Gardens. It was originally linked to the corner of the Quay by a bridge, opposite Duke Street. At the other end of The Quay, Spithead extended into the river for a few yards. Dartmouth sent numerous ships to join the English fleet that attacked the Spanish Armada, including the Roebuck, Crescent and Hart. The Nuestra Señora del Rosario, the Spanish Armada's "payship" commanded by Admiral Pedro de Valdés, was captured along with all its crew by Sir Francis Drake. It was reportedly anchored in the River Dart for more than a year and the crew were used as labourers on the nearby Greenway Estate which was the home of Sir Humphrey Gilbert and his half-brother Sir Walter Raleigh. Greenway was later the home of Dame Agatha Christie. In 1592 the Madre de Deus, a Portuguese treasure ship captured by the English in the Azores, docked at Dartmouth Harbour. It attracted all manner of traders, dealers, cutpurses and thieves and by the time Sir Walter Raleigh arrived to reclaim the Crown's share of the loot, a cargo estimated at half a million pounds had been reduced to £140,000. Still, ten freighters were needed to carry the treasure to London. Henry Hudson put into Dartmouth on his return from North America, and was arrested for sailing under a foreign flag. The Pilgrim Fathers put into Dartmouth's Bayard's Cove, en route from Southampton to America. They rested a while before setting off on their journey in the Mayflower and the Speedwell on 20 August 1620. About 300 miles west of Land's End, upon realising that the Speedwell was unseaworthy, it returned to Plymouth. The Mayflower departed alone to complete the crossing to Cape Cod. Dartmouth's sister city is Dartmouth, Massachusetts. The town contains many medieval and Elizabethan streetscapes and is a patchwork of narrow lanes and stone stairways. A significant number of the historic buildings are listed. One of the most obvious is the Butterwalk, built 1635 to 1640. Its intricately carved wooden fascia is supported on granite columns. Charles II held court in the Butterwalk whilst sheltering from storms in 1671 in a room which now forms part of Dartmouth Museum. Much of the interior survives from that time. The Royal Castle Hotel was built in 1639 on the then new quay. The building was re-fronted in the 19th century, and as the new frontage is itself listed, it is not possible to see the original which lies beneath. A claimant for the oldest building is a former merchant's house in Higher Street, now a Good Beer Guide listed public house called the Cherub, built circa 1380. Agincourt House (next to the Lower Ferry) is also 14th century. The remains of a fort at Gallants Bower just outside the town are some of the best preserved remains of a Civil War defensive structure. The fort was built by Royalist occupation forces in c. 1643 to the south east of the town, with a similar fort at Mount Ridley on the opposite slopes of what is now Kingswear. The Parliamentarian General Fairfax attacked from the north in 1646, taking the town and forcing the Royalists to surrender, after which Gallants Bower was demolished. 19th century Before 1671, what is now the town centre was almost entirely tidal mud flats. The New Road (now Victoria Road) was constructed across the bed of the (silted up) Mill Pool and up the Ford valley after 1823. Spithead was extended in 1864 when the Dartmouth and Torbay Railway arrived in Kingswear and a pontoon was constructed, linked to Spithead by a bridge. The railway directors and others formed the Dartmouth Harbour Commissioners. At this time, all the roads in those parts of Dartmouth which were not land reclamations were very narrow. In 1864-7 Higher Street was widened into Southtown and linked to Lower Street, which was also widened, with the northern part renamed Fairfax Place. Some of the buildings were rebuilt further back with decorative frontages. In 1881 the Harbour Commissioners produced a scheme for an embankment or esplanade from near the Lower Ferry to Hardness, across the remains of The Pool, to provide an attraction for tourists and further mooring space. It was completed in 1885 after much disagreement between the Borough, the Commissioners and the Railway (now the Great Western Railway). A new station was also built at this time. The building of the Embankment left a section of river isolated between Spithead and the New Ground, which is known as The Boatfloat, and is linked to the river by a bridge for small vessels under the road. The coming of steam ships led to Dartmouth being used as a bunkering port, with coal being brought in by ship or train. Coal lumpers were members of gangs, who competed to bunker the ships by racing to be first to a ship. This led to the men living as close as possible to the river, and their tenements became grossly overcrowded, with the families living in slum conditions, with up to 15 families in one house, one family to a room. The Royal National Lifeboat Institution opened the Dart Lifeboat Station at the Sand Quay in 1878, but it was closed in 1896. In all this time only one effective rescue was made by the lifeboat. 20th century The area to the north of Ridge Hill was a shallow and muddy bay ("Coombe Mud") with a narrow road running along the shore linking with the Higher Ferry. The mud was a dumping ground for vessels, including a submarine. The reclamation was completed in 1937 by the extension of the Embankment and the reclamation of the mud behind it, which became Coronation Park. In the 1920s, aided by government grants, the council made a start on clearing the slums. This was aided by the decline in the use of coal as a fuel for ships. The slums were demolished, and the inhabitants were rehoused in new houses in the Britannia Avenue area, to the west of the old village or hamlet of Townstal. The process was interrupted by the second world war, but was resumed with the construction of many prefabs, and later more houses. Community facilities were minimal at first, but a central area was reserved for a church, which was used by the Baptists and opened in 1954, together with a speedway track. The latter was later used for housing, but a new community centre was opened nearby, together with a leisure centre, an outdoor swimming pool, and later an indoor pool, and supermarkets. There are also light industrial units. In the latter part of the Second World War the town was a base for American forces and one of the departure points for Utah Beach in the D Day landings. Slipways and harbour improvements were also constructed. Much of the surrounding countryside and notably Slapton Sands was closed to the public while it was used by US troops for practise landings and manoeuvres. Between 1985 and 1990 the Embankment was widened by 6 metres and raised to prevent flooding at spring tides. A tidal lock gate was provided at the Boatfloat bridge, which could be closed at such times. 21st century Dart Lifeboat Station was reopened in 2007, the first time that a lifeboat had been stationed in the town since 1896. It has initially been kept in a temporary building in Coronation Park. In 2010, a fire seriously damaged numerous historical properties in Fairfax Place and Higher Street. Several were Tudor and Grade I or Grade II listed buildings. Governance The town was an ancient borough, incorporated by Edward III, known formally as Clifton-Dartmouth-Hardness, and consisting of the three parishes of St Petrox, St Saviour and Townstal, and incorporating the hamlets of Ford, Old Mill and Norton. It was reformed under the Municipal Corporations Act 1835. The town returned two members of parliament from the 13th century until 1835, after which one Member of Parliament (MP) was elected until the town was disenfranchised in 1868. It remained a municipal borough until 1974, when it was merged into the South Hams district, and became a successor parish of Dartmouth with a town council. Dartmouth Town Council is the lowest of three tiers of local government. It consists of 16 councillors representing the two wards of Clifton and Townstal. At the second tier, Dartmouth forms part of the Dartmouth and Kingswear ward of South Hams District Council, which returns three councillors. At the upper tier of local government Dartmouth and Kingswear Electoral Division elects one member to Devon County Council. Culture and tourism The Port of Dartmouth Royal Regatta takes place annually over three days at the end of August. The event sees the traditional regatta boat races along with markets, fun fairs, community games, musical performances, air displays including the Red Arrows and fireworks. A Royal Navy guard ship is often present at the event. Other cultural events include beer festivals in February and July (the latter in Kingswear), a music festival and an art and craft weekend in June, a food festival in October and a Christmas candlelit event. The Flavel Centre incorporates the public library and performance spaces, featuring films, live music and comedy and exhibitions. Bayard's Cove has been used in several television productions, including The Onedin Line a popular BBC television drama series that ran from 1971 to 1980. Many of the scenes from the BBC's popular series Down to Earth, starring Ricky Tomlinson, were filmed at various locations around the town. Notable tourist attractions include the Dartmouth Royal Naval College, Bayard's Cove Fort, Dartmouth Castle and the Dartmouth Steam Railway which terminates at Kingswear on the opposite bank of the river. Boat cruises to nearby places along the coast (such as Torbay and Start Bay) and up the river (to Totnes, Dittisham and the Greenway Estate) are provided by several companies. The paddlesteamer PS Kingswear Castle returned to the town in 2013. The South West Coast Path National Trail passes through the town, and also through extensive National Trust coastal properties at Little Dartmouth and Brownstone (Kingswear). The Dart Valley Trail starts in Dartmouth, with routes either side of the River Dart as far as Dittisham, and continuing to Totnes via Cornworthy, Tuckenhay and Ashprington. The area has long been well regarded for yachting, and there are extensive marinas at Sandquay, Kingswear and Noss (approximately one mile north of Kingswear). Climate The nearest Met Office weather station is Slapton, about 5 miles south-south west of Dartmouth and a similar distance from the coast. As with the rest of the British Isles and South West England, the area experiences a maritime climate with warm summers and mild winters—this is particularly pronounced due to its position near the coast—extremes range from a record low of just in January 1987 up to a record high of during June 1976. Transport Dartmouth is linked to Kingswear, on the other side of the River Dart, by three ferries. The Higher Ferry and the Lower Ferry are both vehicular ferries. The Passenger Ferry, as its name suggests, carries only passengers, principally to connect with the Dartmouth Steam Railway at Kingswear railway station. The nearest bridge across the Dart is in Totnes, some away by road. The A379 road runs through Dartmouth, linking the town to Slapton and Kingsbridge to the southwest and to Torbay to the east across the Higher Ferry. The A3122 connects Dartmouth to a junction with the A381, and hence to both Totnes and a more direct route to Kingsbridge. Stagecoach South West provides local town bus services and links to Plymouth, Totnes and Exeter, and Kingsbridge. In addition it provides links to the Torbay resorts of Brixham, Paignton and Torquay from Kingswear via the ferry. No railway has ever run to Dartmouth, but the town does have a railway station, opened on 31 March 1890 to replace the original facility on the pontoon, although it is now a restaurant. In fact Dartmouth Railway Station was built while the line from Paignton to the River Dart was still being built, anticipating a bridge across the river being built near the present Greenway Halt. The railway line to Kingswear was opened in 1864. As a result of shortage of capital, a deviation from the original scheme to run the line from Churston to Greenway with a steamer service to Dartmouth was proposed, but defeated in Parliament. It had been suggested that this could, at a later date, be used as a jumping off point for a bridge to the west bank of the Dart and a line direct to Dartmouth. In 1900, a Light Railway scheme was proposed for a crossing of the Dart near Maypool to join another line from Totnes and then proceed to Kingsbridge and Yealmpton, with a branch to Salcombe. This was also defeated by lack of funds. The railway terminated at a station called "Kingswear for Dartmouth" (now on the Dartmouth Steam Railway) and a ferry took passengers across the river to the station at Dartmouth railway station, which had a dedicated pontoon. British Railways formally closed the line to mainline passenger trains in 1973, but it immediately re-opened as a heritage line and has run as one ever since. Education Britannia Royal Naval College The town is home to the Royal Navy's officer training college (Britannia Royal Naval College), where all officers of the Royal Navy and many foreign naval officers are trained. Schools Dartmouth has one primary school—St John the Baptist R.C. Primary School, and one all-through school—Dartmouth Academy—for those aged 3–16. Dartmouth also has a pre-school in the centre of town, established for over 40 years and based in the old Victorian school rooms at South Ford Road. It provides care for 2- to 5-year-olds and is run as a charitable organisation. Sport and leisure Dartmouth has a Non-League football club Dartmouth A.F.C. who play at Long Cross. Dartmouth also hosts the annual "World Indoor Rally Championship", based on slot car racing in the late summer. At the end of August and early September there is the annual Port of Dartmouth Royal Regatta. Since 1905 Dartmouth has had a greenhouse as part of the Royal Avenue Gardens. In May 2013 this building, used for the previous 10 years by Dartmouth in Bloom, a not-for-profit organisation affiliated with Britain in Bloom, was closed as structurally unsound. There are proposals to restore the greenhouse to its prior Edwardian style. Notable residents George Parker Bidder (1806–1878), the civil engineer and calculating prodigy, notable for his work on railways over much of the world, as well as the docks of the East End in the Port of London. Bidder served on the town council, and his expertise was instrumental in draining the area which is now the centre of the town. He also undertook pioneering work with Samuel Lake on steam trawling whilst living in the town. Bidder died at his home at Paradise Point near Warfleet Creek and is buried at nearby Stoke Fleming. Simon Drew (b.1952), a commercially successful cartoonist and illustrator, lives in Dartmouth and runs a shop on Fosse Street. Gordon Onslow Ford (1912–2003), a leading British surrealist painter, attended the Royal Naval College. Sir John Harvey Jones (1924–2008), businessman and television presenter, attended the Royal Naval College. Rachel Kempson (1910–2003), stage and film actress, was born in Dartmouth. She was the wife of Sir Michael Redgrave and mother of Vanessa, Lynn and Corin, and published her autobiography, Life Among the Redgraves, in 1988. Christopher Robin Milne, son of A. A. Milne, after whom the character Christopher Robin in the Winnie-the-Pooh books was named, used to own the Harbour Bookshop. The bookshop closed in September 2011. Thomas Newcomen, the inventor of the atmospheric engine—the first successful steam-powered pumping engine—was born in Dartmouth in 1663. The location of his house in Lower Street is marked with a plaque, although the building itself was demolished (and elements incorporated into local architect Thomas Lidstone's house on Ridge Hill) in the 19th century to make way for a new road which was named after Newcomen. An 18th-century working Newcomen steam engine is on display in the town. Mary Nightingale, ITV newscaster, lived in Dartmouth for much of her childhood. Flora Thompson lived in Above Town between 1928 and 1940, writing Lark Rise and Over to Candleford during this time. The books were later combined into a single volume with Candleford Green and published as Lark Rise to Candleford. She is buried at Longcross Cemetery. Theodore Veale, recipient of the Victoria Cross during the First World War. John L. Wimbush, landscape and portrait painter. John Flavel, Puritan preacher and writer. References External links Charles Oman, "Dartmouth and Kingswear Castles: Twin Dart estuary defenders" Dartmouth WebCam Towns in Devon Port cities and towns in South West England Seaside resorts in England Civil parishes in South Hams
8579
https://en.wikipedia.org/wiki/Detroit%20Tigers
Detroit Tigers
The Detroit Tigers are an American professional baseball team based in Detroit. The Tigers compete in Major League Baseball (MLB) as a member of the American League (AL) Central division. One of the AL's eight charter franchises, the club was founded in Detroit as a member of the minor league Western League in 1894 and is the only Western League team still in its original city. They are also the oldest continuous one name, one city franchise in the AL. Since their establishment as a major league franchise in 1901, the Tigers have won four World Series championships (, , , and ), 11 AL pennants (1907, 1908, 1909, 1934, 1935, 1940, 1945, 1968, 1984, 2006, 2012), and four AL Central division championships (2011, 2012, 2013, and 2014). They also won division titles in 1972, 1984, and 1987 as a member of the AL East. Since 2000, the Tigers have played their home games at Comerica Park in Downtown Detroit. The Tigers constructed Bennett Park at the corner of Michigan Avenue and Trumbull Avenue in Corktown just west of Downtown Detroit and began playing there in 1896. In 1912, the team moved into Navin Field, which was built on the same location. It was expanded in 1938 and renamed Briggs Stadium. It was renamed Tiger Stadium in 1961 and the Tigers played there until 1999. From 1901 to 2022, the Tigers' overall win–loss record is . The franchise's best winning percentage was .656 in 1934, while its worst was .265 in 2003. Franchise history Western League beginnings (1894–1900) The franchise was founded as a member of the reorganized Western League in 1894. They originally played at Boulevard Park, sometimes called League Park. It was located on East Lafayette, then called Champlain Street, between Helen and East Grand Boulevard, near Belle Isle. In 1895, owner George Vanderbeck decided to build Bennett Park at the corner of Michigan and Trumbull Avenues, which would remain the team's base of operations for the next 104 seasons. The first game at The Corner was an exhibition on April 13, 1896. The team, now occasionally called the "Tigers," beat a local semi-pro team, known as the Athletics, by a score of 30–3. The Tigers played their first Western League game at Bennett Park on April 28, 1896, defeating the Columbus Senators 17–2. At the end of the 1897 season, Rube Waddell was loaned to the team to gain professional experience. After being fined, Waddell left Detroit to pitch in Canada. When the Western League renamed itself the American League for 1900, it was still a minor league, but the next year, it broke from the National Agreement and declared itself a major league, openly competing with the National League for players and for fans in four contested cities. For a while, there were rumors of the team relocating to Pittsburgh. However, these rumors were put to rest when the two leagues made peace in 1903 when they signed a new National Agreement. First Major League season (1901) The Tigers were established as a charter member of the now major league American League in 1901. They played their first game as a major league team at home against the Milwaukee Brewers on April 25, 1901, with an estimated 10,000 fans at Bennett Park. After entering the ninth inning behind 13–4, the team staged a dramatic comeback to win 14–13. The team finished third in the eight-team league. The Cobb era (1905–1926) 1905 In 1905, the team acquired 18-year-old Ty Cobb, a fearless player who came to be regarded as one of the greatest players of all time. The addition of Cobb to an already talented team that included Sam Crawford, Hughie Jennings, Bill Donovan and George Mullin quickly yielded results. 1907 American League Champions Behind the hitting of outfielders Ty Cobb (.350) and Sam Crawford (.323), and the pitching of Bill Donovan and Ed Killian (25 wins each), the Tigers went 92–58 to win the AL pennant in 1907 by 1.5 games over the Philadelphia Athletics. They moved on to their first World Series appearance against the Chicago Cubs. Game 1 ended in a rare 3–3 tie, called due to darkness after 12 innings. The Tigers scored only three runs in the succeeding four games, never scoring more than one run in a game, and lost the Series, 4–0. 1908 American League Champions The Tigers won the AL by just a half-game over the 90–64 Cleveland Naps with a 90–63 record. Cobb hit .324, while Sam Crawford hit .311 with 7 home runs, which was enough to lead the league in the "dead ball" era. The Cubs, however, would defeat the Tigers again in the 1908 World Series, this time in five games. This would be the Cubs' last World Championship until 2016. 1909 American League Champions In 1909, Detroit posted a 98–54 season, winning the AL pennant by 3.5 games over the Athletics. Ty Cobb won the batting triple crown in 1909, hitting .377 with 9 home runs (all inside-the-park) and 107 RBIs. He also led the league with 76 stolen bases. George Mullin was the pitching hero, going 29–8 with a 2.22 ERA, while fellow pitcher Ed Willett went 21–10. Mullin's 11–0 start in 1909 was a Tigers record for 104 years, finally being broken by Max Scherzer's 13–0 start in 2013. It was hoped that a new opponent in the 1909 Series, the Pittsburgh Pirates, would yield different results. The Tigers performed better in the Fall Classic, taking Pittsburgh to seven games, but they were blown out 8–0 in the decisive game at Bennett Park. 1910–1914 The Tigers dropped to third place in the American League in 1910 with an 86–68 record. They posted 89 wins in 1911 to finish second, but were still well behind a powerhouse Philadelphia Athletics team that won 101 games. The team sunk to a dismal sixth place in both the 1912 and 1913 seasons. A bright spot in 1912 was George Mullin pitching the franchise's first no-hitter in a 7–0 win over the St. Louis Browns on July 4, his 32nd birthday. Cobb went into the stands in a May 15, 1912, game to attack a fan that was abusing him, and was suspended. Three days later, the Tigers protested the suspension by fielding a team of replacement players against the Philadelphia Athletics. They lost 24–2. During this five-season stretch, Cobb posted batting averages of .383, .420, .409, .390 and .368, winning the AL batting title every year. 1915 In 1915, the Tigers won a then-club record 100 games, but narrowly lost the AL pennant to the Boston Red Sox, who won 101 games. The 1915 Tigers were led by an outfield consisting of Ty Cobb, Sam Crawford, and Bobby Veach that finished #1, #2, and #3 in RBIs and total bases. Cobb also set a stolen base record with 96 steals in 1915 that stood until 1962, when it was broken by Maury Wills. Baseball historian Bill James has ranked the 1915 Tigers outfield as the greatest in the history of baseball. The only team in Tigers' history with a better winning percentage than the 1915 squad was the 1934 team that lost the World Series to the St. Louis Cardinals. 1916–1920 The Tigers dropped to third place in 1916 with an 87–67 record, and would remain mired in the middle of the AL standings the rest of the decade, never winning more than 80 games. In the late teens and into the 1920s, Cobb continued to be the marquee player, though he was pushed by budding star outfielder Harry Heilmann, who went on to hit .342 for his career. Hughie Jennings left the Tigers after the 1920 season, having accumulated 1,131 wins as a manager. This stood as a Tiger record until 1992, when it was broken by Sparky Anderson. Cobb himself took over managerial duties in 1921, but during his six years at the helm, the Tigers topped out at 86 wins and never won a pennant. 1921 In 1921, the Tigers amassed 1,724 hits and a team batting average of .316, the highest team hit total and batting average in AL history. That year, outfielders Harry Heilmann and Ty Cobb finished #1 and #2 in the American League batting race with batting averages of .394 and .389, respectively. The downfall of the 1921 Tigers, however, was the absence of good pitching. The team ERA was 4.40. Without pitching to support the offense, the 1921 Tigers finished in sixth place in the American League at 71–82, 27 games behind the New York Yankees. On August 19, 1921, Cobb collected his 3,000th career hit off Elmer Myers of the Boston Red Sox. Aged 34 at the time, he is still the youngest player to reach that milestone, also reaching it in the fewest at-bats (8,093). 1922–1926 The Tigers continued to field good teams during Ty Cobb's tenure as player-manager, finishing as high as second in 1923, but lack of quality pitching kept them from winning a pennant. Harry Heilmann hit .403 in 1923, becoming the last AL player to top .400 until Ted Williams hit .406 in 1941. In 1925, Heilmann collected six hits in a season-ending doubleheader to win the batting title, finishing at .393 to Tris Speaker's .389. Cobb announced his retirement in November 1926 after 22 seasons with the Tigers, though he would return to play two more seasons with the Philadelphia Athletics. The Tigers break through (1927–1940) 1927–1933 Though the Tigers struggled with mediocre records in the seven years following Cobb's departure, they were building a solid foundation, adding slugging first baseman Hank Greenberg and pitchers Tommy Bridges and Schoolboy Rowe to a lineup that already included second baseman Charlie Gehringer. In 1927, Harry Heilmann flirted with a .400 batting average all year, eventually finishing at .398 and winning his fourth AL batting title. Following the 1933 season, the Tigers added perhaps the final piece of the puzzle, acquiring catcher Mickey Cochrane from the Philadelphia Athletics to serve as player-manager. 1934 American League Champions The Tigers won the 1934 AL pennant with a 101–53 record, at the time a team record for wins, and still the best win percentage (.656) in team history. The Tigers infield (Hank Greenberg and Charlie Gehringer, along with shortstop Billy Rogell and third baseman Marv Owen) accumulated 462 runs during the season, with Gehringer (214 hits, .356 average) leading the way. Schoolboy Rowe led a strong pitching staff, winning 16 straight decisions at one point of the season and finishing with a 24–8 record. The Tigers would fall in the 1934 World Series in seven games to the "Gashouse Gang" St. Louis Cardinals. After winning a tight battle in Game 5 with a 3–1 decision over Dizzy Dean, Detroit took a 3–2 series lead, but would lose the next two games at Navin Field (Tiger Stadium). For the second time in a World Series Game 7, Detroit folded. St. Louis scored seven times in the third inning off starter Elden Auker and a pair of relievers, while Dean baffled the Tiger hitters en route to an 11–0 victory. The final game was marred by an ugly incident. After spiking Tigers third baseman Marv Owen in the sixth inning, Cardinals left fielder Joe Medwick had to be removed from the game for his own safety by Commissioner Kenesaw Mountain Landis after being pelted with debris from angry fans in the large temporary bleacher section in left field. 1935 World Series Champions The Tigers 1935 lineup featured four future Hall of Famers (Hank Greenberg, Mickey Cochrane, Goose Goslin and Charlie Gehringer). Although they did not challenge the 1934 team's 101 wins, their 93–58 record was good enough to give them the AL pennant by three games over the New York Yankees. Greenberg was named AL MVP after hitting .328 and leading the league in home runs (36), extra-base hits (98) and RBIs (168). Incredibly, Greenberg's RBI total was 48 higher than the next closest player (Lou Gehrig, with 120). The Tigers also got strong contributions from Gehringer (.330), Cochrane (.319) and starting pitchers Tommy Bridges (21–10) and Elden Auker (18–7). The Tigers finally won their first World Series, defeating the Chicago Cubs, 4–2. Game 6 concluded with Goslin's dramatic walk-off RBI single, scoring Cochrane for a 4–3 victory. After owner Frank Navin died in the offseason, Walter Briggs Sr. took over control of the team. 1936–1939 Despite being forecast to win the American League pennant again in 1936, the Tigers fell to a distant second place behind the New York Yankees both that season and in 1937. The team fell further down the standings with an 84–70 record in 1938 and an 81–73 record in 1939. Hank Greenberg nevertheless provided some excitement for Tigers fans in 1938 by challenging the single-season home run record held by Babe Ruth (60). He went into the season's final weekend against the Cleveland Indians with 58 home runs, tied with Jimmie Foxx for the most by a right-handed batter at the time, but he failed to homer. During the final week of the 1938 season, the Tigers presciently held out doubts about a pennant in 1939, but figured that 1940 would be their year. 1940 American League Champions In a tight three-team race, the 90–64 Tigers won the 1940 AL pennant by one game over the Cleveland Indians and two games over the New York Yankees. Prior to the season, first baseman Hank Greenberg was persuaded to move to left field to make room for Rudy York, whom the Tigers had deemed no longer suitable to be their catcher. The move proved successful. York hit .316 with 33 home runs and 134 RBIs. Greenberg batted .340 and slammed 41 home runs while driving in 150. Greenberg won his second AL MVP award, becoming the first major leaguer to win the award at two different positions. Charlie Gehringer batted .313 while collecting 101 walks (for a .428 on-base percentage) and scoring 108 runs. Bobo Newsom was the ace of the Tiger pitching staff in 1940, going 21–5 with a 2.83 ERA. An unlikely hero on the mound this season was 30-year-old rookie Floyd Giebell. Making just his third major league start on September 27, Giebell was called upon to pitch the pennant-clinching game against Bob Feller of the Indians. Feller surrendered just three hits, one being a 2-run homer by Rudy York, while Giebell blanked the Tribe for a 2–0 victory. The Tigers lost the 1940 World Series to the Cincinnati Reds in seven games. Despite a heroic effort by Bobo Newsom, the Tigers came up short in the deciding game, losing 2–1. Newsom's father had died in a Cincinnati hotel room after watching his son win Game 1. An inspired Newsom won Game 5 and pitched Game 7 on just one day's rest. This was the third time the Tigers had lost a World Series in a deciding seventh game. The war years (1941–1945) With Hank Greenberg serving in World War II for all or parts of the 1941–1944 seasons, the Tigers struggled to recapture the glory of 1940. They finished no higher than fifth place in 1941–1943, but did manage a second-place finish in 1944, largely on the strength of pitchers Hal Newhouser and Dizzy Trout, who won 29 and 27 games, respectively. Newhouser, who was 29–9 with a 2.22 ERA, won the first of his two consecutive AL MVP awards this season. The Tigers were in first place as late as September 18, but would finish one game behind the St. Louis Browns for the AL pennant. 1945 World Series Champions With the end of World War II and the timely return of Hank Greenberg and others from the military, the Tigers won the AL pennant by just 1.5 games over the Washington Senators with an 88–65 record. Virgil Trucks returned from the U.S. Navy in time to pitch innings of 1-run ball in the pennant-clinching game, with starter Hal Newhouser pitching the final innings in relief. Newhouser won the pitching triple crown, leading the AL in wins (25), ERA (1.81) and strikeouts (212). He became the first pitcher in the history of the AL, and still the only pitcher as of 2021, to win the MVP Award in two consecutive seasons. With Newhouser, Trucks and Dizzy Trout on the mound and Greenberg leading the offense, Detroit responded in a World Series Game 7 for the first time, staking Newhouser to a 5–0 lead before he threw a pitch en route to a 9–3 victory over the Cubs. Because many stars had not yet returned from the military, some baseball scholars have deemed the 1945 World Series to be among the worst-played contests in World Series history. For example, prior to the World Series, Chicago sportswriter Warren Brown was asked who he liked, and he answered, "I don't think either one of them can win." A long drought (1946–1967) 1946–1950 Following their World Series win in 1945, the Tigers continued to have winning records for the remainder of the decade, finishing second in the AL three times, but never winning the pennant. Hal Newhouser had another outstanding season in 1946, again leading the league in wins (26) and ERA (1.94) while striking out a career-high 275 batters. He nearly won his third straight AL MVP award, finishing second to Ted Williams, who had led the Boston Red Sox to 104 wins (12 games ahead of the second-place Tigers). Also in 1946, the Tigers acquired George Kell, a third baseman who would become a 10-time all-star and Hall of Famer. He batted over .300 in eight straight seasons (1946–53), and finished with a career .306 average. Kell won the batting title in a very close race with Ted Williams in 1949, going 2-for-3 on the last day of the season to edge out the Red Sox slugger, .34291 to .34276. The 1950 season was particularly frustrating, as the Tigers posted a 95–59 record for a .617 winning percentage, the fourth-best in team history at the time. However, they finished that season three games behind a strong New York Yankees team that went on to sweep the Philadelphia Phillies in the World Series. 1951–1960 Over the next 10 years, the Tigers sank to the middle and lower ranks of the American League. The team had only three winning records over this span and never finished higher than fourth place. The last place 1952 team went 50–104 (.325), which was the worst season in Tigers history until the 2003 team lost 119 games. Despite the dismal season, starter Virgil Trucks threw two no-hitters in 1952, becoming only the third pitcher in major league history to accomplished this feat. 1952 also saw Tiger first baseman Walt Dropo get a hit in 12 consecutive plate appearances over a three-game stretch from July 14 to July 15, tying a major league record set by Johnny Kling in 1902. Team owner Walter Briggs Sr. died in 1952. His son Walter Briggs Jr. inherited the team, but he was forced to sell it in 1956 to broadcast media owners John Fetzer and Fred Knorr. Notwithstanding Detroit's fall in the standings, the decade saw the debut of outfielder Al Kaline in 1953. One of the few major league players who never played a day in the minor leagues, he would hit over .300 nine times in his career. He also made 15 All-Star teams, won 10 Gold Gloves, and featured one of the league's best arms in right field. In 1955, the 20-year-old Kaline hit .340 to become the youngest-ever batting champion in major league history. 1958 saw the Tigers become the second to last team to integrate their roster when Dominican player Ozzie Virgil Sr. joined the team. Only the Boston Red Sox trailed the Tigers in integrating their roster. 1961 As the American League expanded from 8 to 10 teams, Detroit began its slow ascent back to success with an outstanding 1961 campaign. The Tigers led the majors in runs scored and won 101 games, a whopping 30-game improvement over the 71–83 1960 team, but still finished eight games behind the Yankees. This marked one of the few times in major league history that a team failed to reach the postseason despite winning 100 or more games, though it had happened once before to the Tigers in 1915. First baseman Norm Cash won the batting title with a .361 average, while teammate Al Kaline finished second. Cash never hit over .286 before or after the 1961 season, and would later say of the accomplishment: "It was a freak. Even at the time, I realized that." Cash's plate heroics, which also included 41 home runs and 132 RBI, might have earned him MVP honors were it not for New York's Roger Maris bashing a then record 61 homers the same season. Cash also drew 124 walks for a league-leading .487 on-base percentage. Tigers outfielder Rocky Colavito actually bettered Cash's home run and RBI totals, with 45 and 140, respectively. The 1961 club featured two non-white starters, Bill Bruton and Jake Wood, and later in the 1960s, black players such as Willie Horton, Earl Wilson, and Gates Brown would contribute to Detroit's rise in the standings. 1962–1966 As a strong nucleus developed, Detroit repeatedly posted winning records throughout the 1960s. In 1963, pitchers Mickey Lolich and Denny McLain entered the rotation. Outfielders Willie Horton (1963), Mickey Stanley (1964) and Jim Northrup (1964) would also come aboard around this time. The team managed a third-place finish during a bizarre 1966 season, in which manager Chuck Dressen and acting manager Bob Swift were both forced to resign their posts because of health problems. Thereafter, Frank Skaff took over the managerial reins until the end of the season. Both Dressen and Swift died during the year; Dressen died of a heart attack in August, while Swift died of lung cancer in October. Following the season, the Tigers hired Mayo Smith to be their new manager. 1967 In 1967, the Tigers were involved in one of the closest pennant races in history. Because of rainouts, the Tigers were forced to play back-to-back doubleheaders against the California Angels over the final two days of the season. They needed to sweep the doubleheader on the last day of the season to force a one-game playoff with the Boston Red Sox. The Tigers won the first game, but lost the second, giving the Red Sox the pennant with no playoff. Detroit finished the season at 91–71, one game behind Boston. Starter Earl Wilson, acquired the previous season from the Red Sox, led the Tigers (and the major leagues) with 22 wins and would form a strong 1–2–3 combination with Denny McLain and Mickey Lolich over the next few years. 1968–1972 Glory in '68: 1968 World Series Champions The Tigers finally returned to the World Series in 1968. The team grabbed first place from the Baltimore Orioles on May 10 and would not relinquish the position, clinching the pennant on September 17 and finishing with a 103–59 record. In a year that was marked by dominant pitching, starter Denny McLain went 31–6 (with a 1.96 ERA), the first time a pitcher had won 30 or more games in a season since Dizzy Dean accomplished the feat in 1934; no pitcher has accomplished it since. McLain won the AL MVP and Cy Young Award for his efforts. 1968 World Series In the 1968 World Series, the Tigers met the defending World Series champion St. Louis Cardinals, led by starter Bob Gibson, who had posted a modern-era record 1.12 ERA during the regular season, and speedy outfielder Lou Brock. This was the first time the Tigers and Cardinals had met in the World Series since 1934. The series was predicated with a bold decision by manager Mayo Smith to play center fielder Mickey Stanley at shortstop, replacing the slick fielding but weak hitting of Ray Oyler. Stanley had never played shortstop before, but was a Gold Glover in the outfield and an excellent athlete. Smith played him at short for the final nine games of the regular season and all seven World Series games, with Oyler only appearing as a late-inning defensive replacement. This allowed Smith to play an outfield of Willie Horton, Jim Northrup and Al Kaline in every game. In Game 1, Gibson completely shut down the Detroit lineup, striking out a World Series record 17 batters en route to an easy 4–0 win. However, due in no small part to pitcher Mickey Lolich's victories in Games 2 and 5, the Tigers climbed back into the World Series. Many fans believe the turning point came in the fifth inning of Game 5, with the Tigers down three games to one, and trailing in the game, 3–2. Left fielder Willie Horton made a perfect throw to home plate to nail Lou Brock, who tried to score from second base standing up, as catcher Bill Freehan blocked the plate with his foot. The Tigers came back with three runs in the seventh to win that game, 5–3, and stay alive. The Cardinals would not threaten to score the rest of this game, and scored only two more meaningless runs over the remainder of the series. In Game 6, McLain ensured a Game 7 by notching his only win of the World Series, a 13–1 blowout, despite pitching on only two days' rest. In Game 7 at Busch Memorial Stadium, Lolich, also pitching on two days' rest, faced Gibson. Both men pitched brilliantly, putting zeros up on the scoreboard for much of the game. In the bottom of the sixth inning, the Cardinals looked primed to take the lead as Lou Brock singled to lead off the inning, only to be promptly picked off by Lolich. One out later, Curt Flood followed with another single, and was also picked off by Lolich. In the top of the seventh, an exhausted Gibson finally cracked, giving up two-out singles to Norm Cash and Willie Horton. Jim Northrup then struck the decisive blow, lashing a triple to center field over the head of Flood, who appeared to misjudge how hard the ball was hit. That scored both Cash and Horton; Northrup himself was then brought home by a Bill Freehan double. Detroit added an insurance run in the ninth. A ninth-inning solo home run by Mike Shannon was all the Cardinals could muster against Lolich as the Tigers took the game, 4–1, and the World Series, 4–3. The Tigers became only the third team to ever win the World Series after being down 3–1. For his three victories that propelled the Tigers to the championship, Lolich was named the World Series Most Valuable Player. , Lolich is the last pitcher to have three complete-game victories in a single World Series. 1969–1971 1969 saw further expansion as both leagues realigned into two divisions of six teams, and the Tigers were placed in the American League East. That year, Detroit failed to defend its title, despite Denny McLain having another outstanding season with a 24–9 campaign, earning him his second straight Cy Young Award (co-winner with Baltimore's Mike Cuellar). The Tigers' 90 wins placed them a distant second in the division to a very strong Baltimore Orioles team, which had won 109 games. The Tigers suffered a disappointing 1970 season, finishing fourth in the AL East with a 79–83 record. Following the season, Mayo Smith was let go and was replaced by Billy Martin. In a playing career that was primarily spent with the New York Yankees, Martin played his final games with the Minnesota Twins and stayed in that organization after his retirement. He managed the Twins to an AL West Division title in 1969, but was fired after that season due to rocky relationships with his players, which included a legendary fight with pitcher Dave Boswell in an alley behind Detroit's Lindell AC sports bar. Also during the offseason, Denny McLain, who had been suspended three times and had a 3–5 record, was part of an eight-player deal with the Washington Senators in what would turn out to be a heist for Detroit. The Tigers acquired pitcher Joe Coleman, shortstop Eddie Brinkman and third baseman Aurelio Rodríguez. Martin's Tigers posted 91 wins in 1971. However, they had to settle for a second-place finish behind the Orioles, who won 101 games to take their third straight AL East Division crown. The season was highlighted by Mickey Lolich's 308 strikeouts, which led the AL and is still the single-season record in franchise history. Lolich also won 25 games and posted a 2.92 ERA while throwing an incredible 376 innings and completing 29 of his 45 starts. Coleman paid immediate dividends for Detroit, winning 20 games, while McLain went 10–22 for the Senators and was out of baseball by the following season. 1972 AL East Champions Joe Coleman, Eddie Brinkman and Aurelio Rodríguez all played critical roles in 1972, when the Tigers captured their first AL East division title. Oddities of the schedule due to an early season strike allowed the 86–70 Tigers to win the division by just game. Brinkman was named Tiger of the Year by the Detroit Baseball Writers, despite a .203 batting average, as he committed just 7 errors in 728 chances (.990 fielding percentage). He also had a streak of 72 games and 331 chances without an error during the season, both AL records for a shortstop. Mickey Lolich was his steady self for the Tigers, winning 22 games with a sparkling 2.50 ERA, while Coleman won 19 and had a 2.80 ERA. Starter Woodie Fryman, acquired on August 2, was the final piece of the puzzle as he went 10–3 over the last two months of the regular season and posted a minuscule 2.06 ERA. Fryman was also the winning pitcher in the division-clinching game against the Boston Red Sox, a 3–1 victory on October 3. 1972 ALCS In the 1972 American League Championship Series, Detroit faced the American League West division champion Oakland Athletics, who had become steadily competitive ever since the 1969 realignment. In Game 1 of the ALCS in Oakland, Mickey Lolich, the hero of '68, took the hill and allowed just one run over nine innings. The Athletics' ace, Catfish Hunter, matched Lolich, surrendering only a solo home run to Norm Cash, and the game went into extra innings. Al Kaline hit a solo homer to break a 1–1 tie in the top of the 11th inning, only to be charged with a throwing error on Gonzalo Márquez's game-tying single in the bottom half of the frame that allowed Gene Tenace to score the winning run. Blue Moon Odom shut down Detroit 5–0 in Game 2. The end of Game 2 was marred by an ugly incident in which Tigers reliever Lerrin LaGrow hit A's shortstop and leadoff hitter Bert Campaneris on the ankle with a pitch. An angered Campaneris threw the bat at LaGrow, and LaGrow ducked just in time for the bat to sail over his head. Both benches cleared, and though no punches were thrown, both LaGrow and Campaneris were suspended for the remainder of the series. It was widely believed that Martin had ordered the pitch that hit Campaneris, who had three hits, two stolen bases and two runs scored in the game. As the series shifted to Detroit, the Tigers caught their stride. Joe Coleman held the A's scoreless on seven hits in Game 3, striking out 14 batters in a 3–0 victory for the Tigers. Game 4 was another pitchers' duel between Hunter and Lolich, resulting again in a 1–1 tie at the end of nine innings. Oakland scored two runs in the top of the 10th and put the Tigers down to their last three outs. Detroit pushed two runs across the plate to tie the game before Jim Northrup came through in the clutch again. His single off Dave Hamilton scored Gates Brown to give the Tigers a 4–3 win and even the series at two games apiece. A first-inning run on an RBI ground out from Bill Freehan, set up by a Gene Tenace passed ball that allowed Dick McAuliffe to reach third, gave Detroit an early lead in the deciding fifth and final game in Detroit. Reggie Jackson's steal of home in the second inning tied it up, though Jackson was injured in a collision with Freehan and had to leave the game. Tenace's two-out single to left field plated George Hendrick to give Oakland a 2–1 lead in the fourth inning. The run was controversial to many Tigers fans, as Hendrick was ruled safe at first base two batters prior to the Tenace hit. Hendrick appeared to be out by two steps on a grounder to short, but umpire John Rice ruled that Norm Cash pulled his foot off first base. Replays and photos, however, show that Cash did not pull his foot. Thanks to that play and four innings of scoreless relief from Vida Blue, the A's took the AL pennant and a spot in the World Series. A slow decline (1973–1978) The 1973 season saw the Tigers drop to third place in the division, with an 85–77 record. Joe Coleman posted another 23 wins, but the other Tiger starters had subpar seasons. Willie Horton hit .316, but injuries limited him to just 111 games. Jim Northrup posted the best batting average of his career (.307) but was inexplicably limited to part-time duty (119 games played), which Northrup attributed to an ongoing feud with Billy Martin that had actually started in the 1972 ALCS. Northrup even proclaimed to the press that Martin "took the fun out of the game." Martin did not survive the 1973 season as manager. He was fired that September after ordering his pitchers to throw spitballs (and telling the press that he did so) in protest of opposing Cleveland Indians pitcher Gaylord Perry, whom Martin was convinced was doing the same. Third base coach Joe Schultz served as interim manager for the remainder of the season. A bright spot for the Tigers in 1973 was relief pitcher John Hiller, who marked his first full season since suffering a heart attack in 1971 by collecting a league-leading 38 saves and posting a brilliant 1.44 ERA. Hiller's saves total would stand as a Tiger record until 2000, when it was broken by Todd Jones' 42 saves (Jones' record would later be broken by José Valverde's 49 saves in 2011). After the season, the Tigers hired Ralph Houk to be their new manager. Houk served in that capacity for five full seasons, through the end of the 1978 season. The roster of players who played under Houk were mostly aging veterans from the 1960s, whose performance had slipped from their peak years. The Tigers did not have a winning season from 1974 to 1977, and their 57 wins in the 1975 season was the team's lowest since 1952. Perhaps the biggest signal of decline for the Tigers was the retirement of Kaline following the 1974 season, after he notched his 3,000th career hit. Kaline finished with 3,007 hits and was elected to the Baseball Hall of Fame in his first year of eligibility in 1980. 1976: The Year of The Bird Tiger fans were provided a glimmer of hope when 21-year-old rookie Mark Fidrych made his debut in 1976. Fidrych, known as "The Bird", was a colorful character known for talking to the baseball and other eccentricities. During a game against the Yankees, Graig Nettles responded to Fidrych's antics by talking to his bat. After making an out, he later lamented that his Japanese-made bat did not understand him. Fidrych entered the All-Star break at 9–2 with a 1.78 ERA, and was the starting pitcher for the American League in the All-Star Game played that year in Philadelphia to celebrate the American Bicentennial. He finished the season with a record of 19–9 and an American League-leading ERA of 2.34. Fidrych, the AL Rookie of the Year, was one of the few bright spots that year with the Tigers finishing next to last in the AL East in 1976. Aurelio Rodríguez won the Gold Glove Award in 1976 at third base, snapping a streak in which Hall of Famer Brooks Robinson had won it for 16 consecutive seasons. 1977–1978 Injuries to his knee, and later his arm, drastically limited Fidrych's appearances in 1977–78. Perhaps more important, however, was the talent coming up through the Tigers farm system at the time. Jack Morris, Lance Parrish, Alan Trammell and Lou Whitaker all made their debuts in 1977, and would help the team to 88 wins in 1978, the only winning season under Houk. The "Bless You Boys" era (1979–1987) Houk's immediate successor as Tigers manager in 1979 was Les Moss, but Moss would only last until June of that year. From June 14, 1979 until the end of the 1995 season, the team was managed by George "Sparky" Anderson, one of baseball's winningest managers and winner of two World Series rings as manager of the Cincinnati Reds during their peak as The Big Red Machine. When Anderson joined the Tigers in 1979 and assessed the team's young talent, he boldly predicted that it would be a pennant winner within five years. Acerbic sports anchor Al Ackerman initiated the phrase "Bless You Boys". Originally used as a sarcastic remark, Ackerman's phrase would take on a new meaning in 1984. The Roar of '84: 1984 World Series Champions As in 1968, the Tigers' next World Series season would be preceded by a disappointing second-place finish, as the 1983 Tigers won 92 games to finish six games behind the Baltimore Orioles in the AL East. The first major news of the 1984 season actually came in late 1983, when broadcasting magnate John Fetzer, who had owned the Tigers since 1957, sold the team to Domino's Pizza founder and CEO Tom Monaghan for $53 million. The 1984 team got off to a 9–0 start highlighted by Jack Morris tossing a nationally televised no-hitter against Chicago in the fourth game of the season. They stayed hot for most of the year, posting a 35–5 record over their first forty games and cruising to a franchise-record 104 victories. The Tigers led the division from opening day until the end of the regular season and finished a staggering 15 games ahead of the second-place Toronto Blue Jays. Closer Willie Hernández, acquired from the 1983 NL champion Philadelphia Phillies in the offseason, won both the Cy Young Award and AL MVP, a rarity for a relief pitcher. 1984 ALCS The Tigers faced the Kansas City Royals in the American League Championship Series. In Game 1, Alan Trammell, Lance Parrish and Larry Herndon went deep to crush the Royals 8–1 at Royals Stadium (now Kauffman Stadium). In Game 2, the Tigers scored twice in the 11th inning when Johnny Grubb doubled off Royals closer Dan Quisenberry en route to a 5–3 victory. The Tigers completed the sweep at Tiger Stadium in Game 3. Marty Castillo's third-inning RBI fielder's choice would be all the help Detroit would need. Milt Wilcox outdueled Charlie Leibrandt, and after Hernandez got Darryl Motley to pop out to preserve the 1–0 win, the Tigers were returning to the World Series. 1984 World Series In the NLCS, the San Diego Padres rallied from losing the first two games to overcome the Chicago Cubs and prevent a fifth Cubs-Tigers series. The Tigers would open the 1984 World Series on the road in San Diego. In Game 1, Larry Herndon hit a two-run home run that gave the Tigers a 3–2 lead. Jack Morris pitched a complete game with 2 runs on 8 hits, and Detroit drew first blood. The Padres evened the series the next night despite pitcher Ed Whitson being chased after pitching of an inning and giving up three runs on five Tiger hits. Tigers starter Dan Petry exited the game after innings when Kurt Bevacqua's three-run homer gave San Diego a 5–3 lead they would not relinquish. When the series shifted to the Motor City, the Tigers took command. In Game 3, a two-out rally in the second inning, highlighted by Marty Castillo's 2-run homer, led to four runs and the yanking of Padres starter Tim Lollar after innings. The Padres never recovered, losing 5–2. Eric Show continued the parade of bad outings in Game 4, getting bounced after innings after giving up a pair of 2-run homers to World Series MVP Alan Trammell in his first two at-bats. Trammell's homers held up with the help of another Morris complete game, and the Tigers' 4–2 win gave them a commanding lead in the series. In Game 5, Kirk Gibson's two-run shot in the first inning would be the beginning of another early end for the Padres' starter Mark Thurmond. Although the Padres would pull back even at 3–3, chasing Petry in the fourth inning in the process, the Tigers retook the lead on a Rusty Kuntz sacrifice fly (actually a pop-out to retreating second baseman Alan Wiggins that the speedy Gibson was able to score on), and then went up 5–3 on a solo homer by Parrish. Gibson came to bat in the eighth inning with runners on second and third and the Tigers clinging to a 5–4 lead. A "Sounds of the Game" video made during the Series by MLB Productions captured this moment, and has been played on TV a number of times since then. Padres manager Dick Williams was shown in the dugout flashing four fingers, ordering an intentional walk, before San Diego reliever Goose Gossage summoned him to the mound. Sparky Anderson was seen and heard yelling to Gibson, "He don't want to walk you!", and making a swing-the-bat gesture. As Anderson had suspected, Gossage threw a 1–0 fastball on the inside corner, and Gibson was ready. He launched the pitch into Tiger Stadium's right field upper deck for a three-run homer, effectively clinching the game and the series. Aurelio López pitched innings of relief and retired all seven batters he faced, earning the win. Despite allowing a rare run in the top of the 8th inning, Willie Hernández got the save as Tony Gwynn flew out to Larry Herndon to end the game, sending Detroit into a wild victory celebration. The Tigers led their division wire-to-wire, from opening day and every day thereafter, culminating in the World Series championship. This had not been done in the major leagues since the 1955 Brooklyn Dodgers. With the win, Anderson became the first manager to win the World Series in both leagues. 1987 AL East Champions After a pair of third-place finishes in 1985 and 1986, the 1987 Tigers faced lowered expectations – which seemed to be confirmed by an 11–19 start to the season. However, the team hit its stride thereafter and gradually gained ground on its AL East rivals. This charge was fueled in part by the acquisition of pitcher Doyle Alexander from the Atlanta Braves in exchange for minor league pitcher John Smoltz. Alexander started 11 games for the Tigers, posting a 9–0 record and a 1.53 ERA. Smoltz, a Michigan native, went on to have a long and productive career, mostly with the Braves, and was ultimately inducted into the Baseball Hall of Fame in 2015. Despite the Tigers' great season, they entered September neck-and-neck with the Toronto Blue Jays. The two teams would square off in seven hard-fought games during the final two weeks of the season. All seven games were decided by one run, and in the first six of the seven games, the winning run was scored in the final inning of play. At Exhibition Stadium, the Tigers dropped three in a row to the Blue Jays before winning a dramatic extra-inning showdown. The Tigers entered the final week of the 1987 season 2.5 games behind. After a series against the Baltimore Orioles, the Tigers returned home trailing by a game and swept the Blue Jays. Detroit clinched the division in a 1–0 victory over Toronto in front of 51,000 fans at Tiger Stadium on October 4. Frank Tanana went all nine innings for the complete-game shutout, and outfielder Larry Herndon gave the Tigers their lone run on a second-inning home run. Detroit finished the season two games ahead of Toronto, securing the best record in the majors (98–64). In what would prove to be their last postseason appearance until 2006, the Tigers were upset in the 1987 American League Championship Series by the 85–77 Minnesota Twins (who in turn won the World Series that year) 4–1. The Twins clinched the series in Game 5 at Tiger Stadium, 9–5. A new approach (1988–1995) Despite their 1987 division title victory, the Tigers proved unable to build on their success. The team lost Kirk Gibson to free agency in the offseason, but still spent much of 1988 in first place in the AL East. A late season slump left the team in second place at 88–74, one game behind the Boston Red Sox. In 1989, the team collapsed to a 59–103 record, worst in the majors. The franchise then attempted to rebuild using a power-hitting approach, with sluggers Cecil Fielder, Rob Deer and Mickey Tettleton joining Trammell and Whitaker in the lineup (fitting for the team with the most 200+ home run seasons in baseball history). In 1990, Fielder led the American League with 51 home runs (becoming the first player to hit 50 since George Foster in 1977, and the first AL player since Roger Maris and Mickey Mantle in 1961), and finished second in the voting for AL MVP. He hit 44 home runs and collected 132 RBI in 1991, again finishing second in the AL MVP balloting. Behind the hitting of Fielder and others, the Tigers improved by 20 wins in 1990 (79–83), and posted a winning record in 1991 (84–78). However, the team lacked quality pitching, despite Bill Gullickson's 20 wins in 1991, and its core of key players began to age, setting the franchise up for decline. Their minor league system was largely barren of talent as well, producing only a few everyday players during the 1990s. Adding insult to injury, the Tigers and radio station WJR announced in December 1990 that they were not renewing the contract of long-time Hall of Fame play-by play announcer Ernie Harwell, and that the 1991 season would be Harwell's last with the team. The announcement was met with resounding protests from fans, both in Michigan and around the baseball world. 1992 saw the Tigers win only 75 games, with Fielder being one of the few bright spots as he won the AL RBI title for a third straight season (124). In August 1992, the franchise was sold to Mike Ilitch, the President and CEO of Little Caesars Pizza who also owned the Detroit Red Wings. One of Ilitch's first moves as the new owner was to rehire Ernie Harwell. Late in the season, Sparky Anderson won his 1,132nd game as a Tiger manager, passing Hughie Jennings for the most all-time wins in franchise history. The team also responded with an 85–77 season in 1993, but it would be their last winning season for a number of years. On October 2, 1995, manager Sparky Anderson chose to not only end his career with the Tigers, but retire from baseball altogether. Randy Smith era (1996–2002) From 1994 to 2005, the Tigers did not post a winning record, the longest sub-.500 stretch in franchise history. In 1996, the Tigers lost a then-team record 109 games, under new general manager Randy Smith. The only team in the majors to have a longer stretch without a winning season during this time were the Pittsburgh Pirates, who did not have a winning record in the years spanning 1993 to 2012. The Tigers' best record over this span was 79–83, recorded in 1997 and 2000. In 1998, the Tigers moved from the AL East, where they had been since the divisions were created in 1969, to the AL Central as part of a realignment necessitated by the addition of the expansion Tampa Bay Devil Rays. In 2000, the team left Tiger Stadium in favor of Comerica Park. Soon after it opened, Comerica Park drew criticism for its deep dimensions, which made it difficult to hit home runs; the distance to left-center field (395 ft), in particular, was seen as unfair to hitters. This led to the nickname "Comerica National Park." The team made a successful bid to bring in slugger Juan González from the Texas Rangers for the inaugural season at Comerica Park. After four consecutive seasons of no fewer than 39 home runs, González only hit 22 homers in 2000. He cited Comerica Park's dimensions as a major reason why he turned down a multiyear contract extension. In 2003, the franchise largely quieted the criticism by moving in the left-center fence to , taking the flagpole in that area out of play, a feature carried over from Tiger Stadium. In 2005, the team moved the bullpens to the vacant area beyond the left field fence and filled the previous location with seats. In late 2001, Dave Dombrowski, former general manager of the 1997 World Series champion Florida Marlins, was hired as team president. In 2002, the Tigers started the season 0–6, prompting Dombrowski to fire the unpopular Smith, as well as manager Phil Garner. Dombrowski then took over as general manager and named bench coach Luis Pujols to finish the season as interim manager. The team finished 55–106. After the season was over, Pujols was let go. Most losses in American League history (2003) Dave Dombrowski hired popular former shortstop Alan Trammell to manage the team in 2003. With fellow 1984 teammates Kirk Gibson and Lance Parrish on the coaching staff, the rebuilding process began. On August 30, 2003, the Tigers' defeat at the hands of the Chicago White Sox caused them to join the 1962 New York Mets, who were a first year expansion team, as the only modern MLB teams to lose 100 games before September. They avoided tying the 1962 Mets' modern MLB record of 120 losses only by winning five of their last six games of the season, including three out of four against the Minnesota Twins, who had already clinched the AL Central and were resting their stars. Mike Maroth went 9–21, becoming the first pitcher to lose 20 games in more than 20 years. Maroth, Jeremy Bonderman (6–19), and Nate Cornejo (6–17) were the top three pitchers in losses in the entire major leagues, the first time in history that this had occurred. (The 1952 Tigers had the top three losers in the AL, but not the major leagues.) The Tigers finished 43–119, the worst record in franchise history. This eclipsed the previous AL record of 117 losses set by the 1916 Philadelphia Athletics. While the 2003 Tigers rank as the third worst team in major league history based on total losses, they fare slightly better based on winning percentage. Their .265 win percentage is the majors' sixth-worst since 1900. Rebuilding the franchise (2004–2006) After the embarrassing 2003 season, the Tigers vowed to make changes. Under Dave Dombrowski, the franchise demonstrated a willingness to sign marquee free agents. In 2004, the team signed or traded for several talented but high-risk veterans, such as Fernando Viña, Rondell White, Iván Rodríguez, Ugueth Urbina, and Carlos Guillén, and the gamble paid off. The 2004 Tigers finished 72–90, a 29-game improvement over the previous season. This was the largest improvement in the AL since the Baltimore Orioles had a 33-game improvement from 1988 to 1989. Prior to the 2005 season, the Tigers spent a large sum for two prized free agents, Troy Percival and Magglio Ordóñez. On June 8, 2005, the Tigers traded pitcher Ugueth Urbina and infielder Ramón Martínez to the Philadelphia Phillies for Plácido Polanco. The Tigers stayed on the fringes of contention for the AL wild card for the first four months of the season, but then faded badly, finishing 71–91. The collapse was perceived as being due both to injuries and to a lack of player unity; Rodríguez in particular was disgruntled, taking a leave of absence during the season to deal with a difficult divorce. Trammell, though popular with the fans, took part of the blame for the poor clubhouse atmosphere and lack of continued improvement, and he was fired at the end of the season. A highlight of the 2005 campaign was Detroit's hosting of the Major League Baseball All-Star Game, its first since 1971. In the Home Run Derby, Rodríguez finished second, losing to the Phillies' Bobby Abreu. In October 2005, Jim Leyland, who managed Dombrowski's 1997 World Series champion Florida Marlins, replaced Trammell as manager; two months later, in response to Troy Percival's arm problems, closer Todd Jones, who had spent five seasons in Detroit (1997–2001), signed a two-year deal to return to the Tigers. Veteran left-hander Kenny Rogers also joined the Tigers from the Texas Rangers in late 2005. The return of the Tigers: 2006 American League Champions After years of futility, the 2006 season showed signs of hope. The impressive rookie campaigns of eventual AL Rookie of the Year Justin Verlander, centerfielder Curtis Granderson, and flamethrowing relief pitcher Joel Zumaya, coupled with a well-publicized early-season tirade by Leyland, helped the team explode and quickly rise to the top of the AL Central. The team reached a high point when they were 40 games over .500, but a second half swoon started to raise questions about the team's staying power. On August 27, a 7–1 victory over the Cleveland Indians gave the Tigers their 82nd victory and their first winning season since 1993. On September 24, the Tigers beat the Kansas City Royals 11–4 to clinch their first playoff berth since 1987. A division title seemed inevitable. All that was required was one win in the final five games of the season, which included three games against the Royals, whom the Tigers had manhandled much of the season. However, the Tigers lost all five games to finish 95–67, and the division title went to the 96–66 Minnesota Twins. The Tigers instead settled for the AL wild card. The playoffs saw the Tigers beat the heavily favored New York Yankees 3–1 in the ALDS and sweep the Oakland Athletics in the 2006 ALCS, thanks to a walk-off home run in Game 4 by right fielder Magglio Ordóñez. They advanced to the World Series, where they lost to the underdog St. Louis Cardinals in five games. Falling short (2007–2010) 2007 During the offseason, the Tigers traded for outfielder Gary Sheffield, who had been a part of the 1997 World Series champion Florida Marlins managed by Jim Leyland. In addition to acquisitions, Dombrowski developed a productive farm system. Justin Verlander and Joel Zumaya, the most notable rookie contributors to the 2006 team, were followed by Andrew Miller, who was drafted in 2006 and called up early in the 2007 campaign, and minor leaguer Cameron Maybin, an athletic five-tool outfielder ranked #6 in Baseball America's 2007 Top 100 Prospects. On June 12, Verlander threw the Tigers' first no-hitter since 1984 (Jack Morris) and the first in Comerica Park history, in a 4–0 win over the Milwaukee Brewers. The Tigers had the best record in baseball in late July, but lost a few players to injuries and started to play poorly in the second half. The Tigers were officially eliminated from playoff competition on September 26, 2007 when the New York Yankees clinched a wild card berth. The Tigers, at 88–74, finished second in the AL Central. Magglio Ordóñez captured the AL batting title in 2007 with a .363 average, becoming the first Tiger to win it since Norm Cash did so in 1961. 2008 Going into the 2008 season, the franchise traded for prominent talent in Édgar Rentería (from the Atlanta Braves) and Miguel Cabrera and Dontrelle Willis (from the Florida Marlins). However, the Tigers, who now boasted the second-highest team payroll in the majors at over $138 million, began the regular season by losing seven straight games. The Tigers climbed back, and at the midway point of the season, they were 42–40. In the end, the team finished miserably, slumping to a 74–88 record. Justin Verlander finished with his worst season as a pro, as he went 11–17 with a 4.84 ERA. The Tigers also lost closer Todd Jones to retirement on September 25, 2008. Despite the disappointing season, the team set an attendance record in 2008, drawing 3,202,654 customers to Comerica Park. 2009 Going into the 2009 season, the Tigers acquired starter Edwin Jackson from the 2008 AL Champion Tampa Bay Rays, and called up rookie and former #1 draft pick Rick Porcello. Jackson was outstanding in the first half, making his first All-Star team, while Porcello was solid most of the year, posting a 14–9 record with a 3.96 ERA and displaying grit and maturity beyond his 20 years of age. Justin Verlander bounced back from an off 2008 to win 19 games. He posted a 3.45 ERA and led the AL in strikeouts (269) to finish third in the AL Cy Young balloting. Fernando Rodney assumed the closer role in spring training, replacing the retired Todd Jones. Rodney responded with 37 saves in 38 tries, while Bobby Seay, Fu-Te Ni, Brandon Lyon, and Ryan Perry shored up the middle relief that plagued the team in 2007 and 2008. Despite the improvements, the Tigers once again found themselves struggling to hold a lead in the AL Central. The team entered September with a 7-game lead in the division, but wound up tied with the Minnesota Twins at 86 wins by the final day of the regular season. The season ended on October 6 with a 6–5 loss in 12 innings to the Twins in the tie-breaker game, leaving the Tigers with an 86–77 record. The Tigers spent 146 days of the 2009 season in first place, but became the first team in Major League history to lose a three-game lead with four games left to play. 2010 Entering 2010, the Tigers parted ways with Curtis Granderson and Edwin Jackson as part of a three-way trade with the New York Yankees and Arizona Diamondbacks; in return they picked up outfield prospect Austin Jackson and pitchers Phil Coke, Max Scherzer and Daniel Schlereth. Jackson made the Tigers opening day roster, and was American League Rookie of the Month for April. 2010 also saw the debut of Brennan Boesch, who was named the AL Rookie of the Month for May and June. At the All-Star break, the Tigers were a half-game out of first place in the AL Central, behind the Chicago White Sox. However, a slow start after the break and injuries to three key players sent the Tigers into yet another second half tailspin. The Tigers finished the season in third place with an 81–81 record, 13 games back of the division-winning Minnesota Twins. While playing outstanding baseball at home, the Tigers were just 29–52 on the road. Among the season highlights were Miguel Cabrera hitting .328 with 38 home runs and an AL-best 126 RBI, earning the AL Silver Slugger Award at first base and finishing second in the AL MVP race (earning 5 of 28 first-place votes). Jackson (.293 average, 103 runs, 181 hits, 27 stolen bases) finished second in the AL Rookie of the Year voting. Justin Verlander enjoyed another strong season (18–9 record, 3.37 ERA, 219 strikeouts). The near-perfect game On June 2, 2010, Armando Galarraga was pitching a perfect game against the Cleveland Indians with 2 outs in the top of the ninth inning when first-base umpire Jim Joyce made a controversial call, ruling Jason Donald safe at first. Video replay showed he was out. A tearful Joyce later said, "I just cost that kid a perfect game. I thought he beat the throw. I was convinced he beat the throw, until I saw the replay." Galarraga would later tell reporters that Joyce apologized to him directly and gave him a hug. The next day, with Joyce umpiring home plate, Galarraga brought out the lineup card and the two shook hands. Despite nationwide support for overturning the call, which included supportive statements from the Governor of Michigan and the White House, commissioner Bud Selig let the call stand. However, he said he would look into expanding instant replay in the future. Division winners (2011–2014) 2011: First AL Central Championship The Tigers returned much of their roster from 2010, while adding relief pitcher Joaquín Benoit, catcher/DH Victor Martinez, and starting pitcher Brad Penny. On May 7, Verlander took a perfect game against the Toronto Blue Jays into the 8th inning. After a walk to J. P. Arencibia, Verlander coaxed a double-play grounder and went on to the 9th inning to complete his second career no-hitter by facing the minimum 27 batters. It was the seventh no-hitter in Tigers history. On August 27, Verlander defeated the Minnesota Twins, 6–4, to become the first Tiger since Bill Gullickson in 1991 to win 20 games in a season. Verlander also became the first major league pitcher since Curt Schilling in 2002 to reach 20 wins before the end of August. In May, the Tigers were as many as eight games back of the first place Cleveland Indians. However, they would start to play better. The Tigers sent five players to the 2011 All-Star Game. Catcher Alex Avila was voted in as a starter, while Justin Verlander, José Valverde and Miguel Cabrera were added as reserves. Verlander was unavailable to play in the All-Star Game due to the rule where starting pitchers who play the Sunday beforehand are ineligible. Shortstop Jhonny Peralta was later added to the All-Star team when the Yankees' Derek Jeter was unable to play due to injury. As a three-way battle for the division title developed between the Tigers, Indians, and Chicago White Sox, the Tigers put together an 18–10 record in August to begin to pull away. Starter Doug Fister, who was acquired at the trade deadline, provided an immediate spark, going 8–1 over the final two months of the season with a sparkling 1.79 ERA. After a loss on September 1, the Tigers reeled off a 12-game winning streak to put any thoughts of another late-season collapse to rest. The streak consisted of four consecutive three-game sweeps over their AL Central Division rivals. It was the Tigers' longest winning streak since the 1934 team won 14 straight. On September 16, the Tigers clinched the AL Central Division title with a 3–1 win over the Oakland Athletics. It was their first AL Central title since joining the division in 1998, and first division title of any kind since 1987. Members of the 2011 Tigers won multiple statistical awards in 2011. Verlander won the pitching triple crown, leading the AL in wins (24), ERA (2.40) and strikeouts (250). On November 15, Verlander was a unanimous selection for the AL Cy Young Award. In a much closer vote six days later, Verlander also won AL MVP, becoming the first pitcher to do so since Dennis Eckersley in 1992. Valverde was the AL saves leader with 49 (in 49 save opportunities), winning the 2011 MLB Delivery Man of the Year Award. Cabrera won the AL batting title with a .344 average, while also leading the AL in on-base percentage (.448) and doubles (48). The Tigers beat the New York Yankees by a score of 3–2 in Game 5 of the ALDS, winning the series 3–2. They advanced to the ALCS, but they lost to the defending AL Champion Texas Rangers, 4–2. 2012: American League Champions In 2012, the Tigers looked to defend their 2011 AL Central Division title, with the hopes of earning a second consecutive playoff appearance. On January 24, the Tigers signed free agent All-Star first baseman Prince Fielder to a 9-year, $214 million contract. The move came shortly after the Tigers learned that Víctor Martínez had torn his anterior cruciate ligament during offseason training in Lakeland, Florida, and would likely miss the entire 2012 season. Miguel Cabrera moved back to his original position of third base, leading to the eventual release of veteran Brandon Inge on April 26. On July 23, the Tigers acquired veteran second baseman Omar Infante, who played for Detroit in 2003–07, and starting pitcher Aníbal Sánchez from the Miami Marlins in exchange for starting pitcher Jacob Turner and two other minor leaguers. At the midway point of the 2012 season, the Tigers were three games under .500 (39–42). The team played much better in the second half and, after a fierce battle down the stretch with the Chicago White Sox, the Tigers clinched the AL Central division title on October 1 with a 6–3 win against the Kansas City Royals. Coupled with the Tigers' division title in 2011, it marked the first back-to-back divisional titles in team history, and first back-to-back postseason appearances since 1934–35. The Tigers concluded the season with an 88–74 record. On the final day of the season, Cabrera earned the Triple Crown in batting, leading the AL in batting average (.330), home runs (44), and runs batted in (139). No player had accomplished this feat since Carl Yastrzemski in 1967. On the mound, starters Justin Verlander and Max Scherzer finished first and second among the American League strikeout leaders, with 239 and 231, respectively. Verlander (17–8, 2.64 ERA) finished second in the Cy Young Award balloting to David Price of the Tampa Bay Rays. In the American League Division Series, the Tigers defeated the Oakland Athletics, 3–2, earning their second straight trip to the American League Championship Series. The Tigers completed a four-game sweep of the New York Yankees in the ALCS to win their 11th AL pennant and earn a trip to the World Series. In the World Series, the Tigers were swept by the San Francisco Giants. On November 15, 2012, Cabrera was named AL MVP. 2013: American League title defense The Tigers entered the 2013 season looking to defend their 2012 AL pennant. Key acquisitions in the offseason included signing free agent outfielder Torii Hunter to a two-year, $26 million contract, while also signing their 2012 trade deadline acquisition, pitcher Aníbal Sánchez, to a five-year, $80 million deal. The Tigers also signed free agent catcher Brayan Peña to a one-year contract. Moreover, ace starter Justin Verlander signed a $180 million contract extension. The Tigers placed six players on the 2013 American League All-Star team: Miguel Cabrera, Prince Fielder, Jhonny Peralta, Torii Hunter, Max Scherzer and Justin Verlander. On September 25, the Tigers clinched their third consecutive AL Central Division title. Tigers pitchers struck out 1,428 batters during the regular season, breaking the record of 1,404 held by the 2003 Chicago Cubs. Cabrera (.348 average, 44 HR, 139 RBI) was voted the AL MVP for the second straight season, while Scherzer (21–3, 2.90 ERA, 240 strikeouts) won the AL Cy Young Award. The Tigers played the Oakland Athletics for the second straight year in the ALDS, and defeated the A's, 3–2. The Tigers set a record by striking out 57 Oakland batters in the ALDS. With his Game 5 gem, Verlander ran his postseason scoreless streak against Oakland to 30 innings. The Tigers advanced to their third straight ALCS, where they played the Boston Red Sox for the first ever time in the postseason. The Tigers would be defeated in six games. 2014: Changes at the top Jim Leyland stepped down from his managerial position after eight years with Detroit, and the Tigers hired Brad Ausmus as Leyland's successor. On November 20, 2013, the Tigers traded Prince Fielder to the Texas Rangers for three time All-Star second baseman Ian Kinsler plus cash considerations with regard to Fielder's remaining contract amount. The Tigers later traded starting pitcher Doug Fister to the Washington Nationals for infielder Steve Lombardozzi Jr. and pitchers Ian Krol and Robbie Ray. In an effort to improve a bullpen that often struggled in 2013, the Tigers signed veteran closer Joe Nathan to a two-year, $20 million contract, with a club option for 2016, and later signed Joba Chamberlain to a one-year, $2.5 million deal. On May 2, 2014, a month into the season, with the bullpen having a combined 5.37 ERA, which was 29th out of 30 in MLB, the Tigers signed free agent reliever Joel Hanrahan to a one-year contract. However, he never came off the disabled list to pitch for the team. The team further bolstered the bullpen near the MLB trading deadline, dealing pitchers Corey Knebel and Jake Thompson to the Texas Rangers in exchange for former All-Star closer Joakim Soria on July 23. On July 31, with just hours left before the end of the non-waiver trade deadline, the Tigers traded pitcher Drew Smyly and shortstop Willy Adames to the Tampa Bay Rays, and Austin Jackson to the Seattle Mariners in a three-team deal to acquire pitcher David Price from the Rays. With the acquisition of Price, the Tigers became the first team in major league history with three consecutive Cy Young Award winners in its starting rotation. On September 28, the last day of the regular season, Price pitched a 3–0 gem against the Minnesota Twins, and the Tigers clinched their fourth consecutive AL Central Division title. The 90–72 Tigers finished one game ahead of the Kansas City Royals. The Tigers faced the Baltimore Orioles in the 2014 American League Division Series, where they were swept, 3–0. J. D. Martinez became the first player in franchise history to hit home runs in his first two career postseason games. Both were part of back-to-back homers, with Víctor Martínez and Nick Castellanos in Games 1 and 2, respectively. 2015: Mid-season Reboot Brad Ausmus continued to manage the Tigers for a second season. Free agents Max Scherzer and Torii Hunter left for other teams at the end of the year, while Rick Porcello, Eugenio Suárez, Robbie Ray, and prospect Devon Travis were all lost through trades. On the receiving end, the Tigers traded for slugger Yoenis Céspedes, relief pitcher Alex Wilson, speedy outfielder Anthony Gose and starting pitchers Alfredo Simón and Shane Greene. After winning the first six games of the year in record-breaking fashion, the season slowly went downhill for the Tigers. Inconsistent pitching, division rivals outperforming expectations, and injuries to multiple players, including Joe Nathan, who only appeared in one game, Victor Martínez, and career first stints on the disabled list for Justin Verlander and Miguel Cabrera, sent the team below the .500 mark as the trade deadline came and the decision was made to "reboot" the team. Within a two-day span in late July, the Tigers traded David Price, Joakim Soria and Cespedes, receiving six well regarded prospects in return, including Daniel Norris and Michael Fulmer. On August 4, longtime general manager Dave Dombrowski was released by the team, with assistant GM Al Avila being promoted to general manager and president of baseball operations. Despite difficulties, the Tigers still ended up sending four players to the 2015 MLB All-Star Game: Miguel Cabrera received his 10th career All-Star selection and the starting nod, but could not play due to injury; David Price received his 5th career selection as well as the credit for the win for the American League; and J.D. Martinez and Jose Iglesias both received their first career All-Star selections. The Tigers ended the season in last place in the AL Central Division with a record of 74–87. The pitching staff was one of the worst in MLB, ending 27th in ERA, 28th in FIP, and 27th in WHIP. However, the team also ended with a team batting average of .270, the best in MLB, while Miguel Cabrera finished with the highest player batting average in the AL and MLB (.338), earning his fourth batting title in five years. 2016 Brad Ausmus and the Tigers entered the 2016 season with new pitching coach Rich Dubee and 9 out of 25 members of the 2015 roster being replaced through trades and free agency. Prominent additions included two highly sought free agents, starting pitcher Jordan Zimmermann and outfielder Justin Upton, as well as players acquired through trades: outfielder Cameron Maybin, and the veteran closer Francisco Rodriguez, who led a totally revamped bullpen. The Tigers lost two 2016 draft picks due to free agent compensation but, because of their bottom-ten finish in 2015, they kept their first round pick. Key veteran losses include catcher Alex Avila and outfielder Rajai Davis, who both signed free agent deals with other teams in the division. The Tigers finished the season with a record of 86–75, eight games behind the first place Cleveland Indians. Detroit was the final team to fall out of contention for a wild card spot, losing Saturday and Sunday games to the Atlanta Braves, while the two teams they were chasing, Baltimore and Toronto, got needed wins. Pitcher Michael Fulmer, acquired from the New York Mets organization in 2015, won the 2016 AL Rookie of the Year Award. Rebuilding (2017–present) 2017 Mike Ilitch, the Tigers owner since 1992, died at the age of 87 on February 10, 2017. The team remains in an Ilitch family trust, under the leadership of Mike's son, Christopher Ilitch. Like the Detroit Red Wings, the Tigers honored their owner in multiple ways, the most prominent being a "Mr. I" uniform patch. After a disappointing record through the All-Star break, the Tigers began committing to a rebuild, trading J. D. Martinez, Alex Avila and Justin Wilson in July, plus Justin Upton and Justin Verlander in August. On September 22, the Tigers announced that the team would not extend manager Brad Ausmus' contract past the 2017 season, ending his four-year tenure as manager. Under the management of Ausmus, the Tigers had a record of 314–332 (.486 winning percentage) and won one AL Central division title in 2014. The Tigers went 6–24 in September, ending the season in a tie for the worst record in MLB with the San Francisco Giants. However, due to a tiebreaker, the Tigers were awarded the number one overall pick in the 2018 MLB Draft. 2018 On October 20, 2017, the Tigers announced that Ron Gardenhire and the team had reached a three-year agreement for the former Minnesota Twins skipper to succeed Brad Ausmus as the team's manager. While rebuilding with young players, the team also lost slugger Miguel Cabrera and expected ace pitcher Michael Fulmer to extensive stints on the disabled list. This led the team to the same 64–98 record as the previous year, the fifth worst record in MLB, but still good for third place in a very weak AL Central division. 2019 Significant changes to the 2019 season opening roster include the departure of José Iglesias, and the one-year signings of the middle infield tandem of Jordy Mercer and Josh Harrison, both formerly of the Pittsburgh Pirates. On March 19, it was announced that Michael Fulmer would undergo Tommy John surgery and miss the 2019 season. The Tigers finished the season with a 47–114 record, the worst in all of Major League Baseball, and their second worst season in franchise history after their 43–119 record in 2003. The 2019 Tigers tied the 1939 St. Louis Browns for the most home losses (59) during a season in the modern era. 2020 On April 6, Al Kaline died at the age of 85. Kaline had been affiliated with the team for 67 years, most recently as an executive. The Tigers wore a No. 6 patch to honor him. In the shortened 60-game season due to the COVID-19 pandemic, the Tigers remained in playoff contention heading into September. On September 19, manager Ron Gardenhire announced his retirement from baseball effective immediately, citing health concerns. Bench coach Lloyd McClendon was named interim manager for the remainder of the season. The Tigers ultimately faded down the stretch, finishing with a 23–35 record. 2021 On October 30, 2020, the Tigers hired A. J. Hinch as manager. On May 18, 2021, Spencer Turnbull pitched the eighth no-hitter in Tigers history against the Seattle Mariners. Turnbull became the first Tigers player to throw a no-hitter since Justin Verlander on May 7, 2011. The Tigers finished the season with a 77–85 record. This outperformed preseason predictions for the team, most of which said the Tigers would finish last. 2022 On August 10, 2022, the Tigers fired Al Avila as general manager. On September 19, 2022, the Tigers hired Scott Harris as president of baseball operations. The Tigers finished the season with a 66–96 record and had one of the worst offense in the league. On October 25, 2022, the Tigers hired Rob Metzler as vice president and assistant general manager. Best seasons in Detroit Tigers history Worst seasons in Detroit Tigers history Payroll Nickname There are various legends about how the Tigers got their nickname. One involves the striped socks they wore. George Stallings, the team's manager, took credit for the name. However, the earliest known use of it appeared in the Detroit Free Press on April 16, 1895, a year prior to Stallings joining the team. In the book A Place for Summer: A Narrative History of Tiger Stadium, Richard Bak states that the name originated from the Detroit Light Guard military unit, who were known as "The Tigers". They had played significant roles in certain Civil War battles and in the 1898 Spanish–American War. Upon entry into the majors, the ballclub sought and received formal permission from the Light Guard to use its trademark. From that day forth, the team has been known officially as the Tigers. Rivalries and fan base The Tigers' rivalries with other baseball franchises have changed throughout the years, with no one rivalry standing out. The most notable of them are with regional neighbors Cleveland Guardians (formerly Indians) and Chicago White Sox. The others are with nearby teams such as the Kansas City Royals, Minnesota Twins, and the Toronto Blue Jays, the latter a holdover from when the Tigers competed in the AL East. There are numerous Tigers fans along the Lake Erie region throughout the state of Michigan, northwestern Ohio, southwestern Ontario, as well as a small fan base in and around the Erie, Pennsylvania area, due in part to Detroit's proximity to these regions as well as the presence of the Tigers' Double-A affiliate Erie SeaWolves in northwestern Pennsylvania. The Tigers have their Triple-A affiliate Toledo Mud Hens in Toledo, Ohio in addition to their Double-A affiliate in Erie. The cities of Windsor and Sarnia, Ontario, have large fan bases of loyal Tigers fans. The Tigers continue to develop a strong and long line of baseball fans in Ontario; the majority of baseball fans in southwestern Ontario are considered Tigers loyalists. The Tigers have had some rivalries with NL teams that they have faced repeatedly in the World Series, such as the Chicago Cubs (four times) and St. Louis Cardinals (three times). In interleague play, the Pittsburgh Pirates are the Tigers' "natural rival." The rivalry with Cleveland came to a head when the Tigers played at Progressive Field on August 7, 2013, with the teams first and second in the AL Central standings. Many Tigers fans who made the short trip to Cleveland started several "Let's go Tigers!" chants while the game was tied in the 9th inning. Cleveland fans combated this with a "Detroit's bankrupt!" chant, in reference to the city's 2013 bankruptcy. Footage of the game from SportsTime Ohio that had the chants clearly audible quickly went viral, with many baseball fans on social media criticizing Indians fans for the chant due to the circumstances of Detroit's financial situation. The Tigers ended up defeating Cleveland 6–5 in 14 innings. Home attendance Comerica Park Rally cry During the 1968 season, the team was cheered on by the phrase, "Go Get 'Em, Tigers", which was made popular by a song of the same name written and recorded by Artie Fields. "Sock it to 'em, Tigers!" was also a popular phrase during this time. During the 1984 World Series championship run, the team was cheered on to the cry, "Bless You Boys," a phrase coined by sportscaster Al Ackerman. In 2005, the team began using the phrase "Who's Your Tiger?" as its slogan. During the 2006 season, a rally cry caught on in the Tigers' dugout. In a June game versus the New York Yankees, Tigers pitcher Nate Robertson was featured in an in-game interview with FSN Detroit. During the interview, Robertson began to stuff Big League Chew bubble gum into his mouth. The Tigers would spark a rally, in which they came back to tie the game. As a result, the phrase "Gum Time" became popular. Additionally, the chant of a local man, the late James Van Horn, who patrolled the streets around Comerica Park yelling out "Eat 'Em Up Tigers! Eat 'Em Up!", was very popular. In 2009, the team used the phrase "Always a Tiger" as its slogan. In 2011, the slogan was switched back to "Who's Your Tiger?" During the 2018 season, a rally cry caught on with both players and fans. In a May game versus the Los Angeles Angels, a goose appeared on the field during a second rain delay. After a few minutes, it left the field only to fly into the scoreboard. The goose was unharmed. When the game resumed, the Tigers scored five runs in the sixth inning. They ultimately won the game, 6–1. As a result, the goose was referred to as the "Rally Goose" and the phrase became popular. Uniforms and logos The Tigers have worn essentially the same home uniform since 1934 — solid white jersey with navy blue piping down the front and an Old English "D" on the left chest, white pants, navy blue hat with a white letter D in the blackletter or textur/textualis typeface associated with Middle and Early Modern English and popularly referred to as "Old English" even though it was not used for that language. On the Tigers' road uniforms, the D on their hats is orange and a script "Detroit" appears across the jersey. A version of the team's blackletter D was first seen on Tigers uniforms in 1904, after using a simple block D in 1903. The blackletter D appeared frequently after that until being established in 1934. In 1960, the Tigers changed their uniform to read "Tigers" in script form, but the change only lasted one season before the traditional uniform was reinstated. The Tigers used to have different versions of the Old English D on the cap and jersey. In 2018, the Tigers changed the classic curved logo on their home uniforms to match that of the sharp-cornered hat logo; additionally, the logo on the cap was enlarged. In 2019, the Tigers reverted to the smaller logo on the cap, but kept the cap curved 'D' on the home jerseys. In 2023, the Tigers entered into a uniform sponsorship deal with Michigan-based supermarket chain Meijer. The patch, which is located on either sleeve depending on the player's handedness, matches the colors of the team's uniforms. Unique characteristics of the Tigers uniforms: The Tigers are the only team in Major League Baseball to have a color on their road uniforms that is not on their home uniforms (orange). The Tigers first added orange to their road uniform in 1903, but it was not until 1972 that the Tigers made orange an accent color on their road uniforms full-time. The Tigers also wore the "orange D" navy caps at home during the late 1940s through the 1950s. The Tigers' uniforms have more belt loops than those of any other team, owing to the fact that their uniform pants do not feature the wide "tunnel" loops that appear on most baseball pants. Alternate jerseys: In 1995, the Tigers introduced a solid navy blue alternate jersey, which featured the team's primary logo at the time, a tiger stepping through the D, on the chest. It was worn for one home game. The Tigers wear the throwback jerseys of the Detroit Stars for their annual Negro Leagues Tribute Game in order to pay tribute to the Negro leagues players and their contributions to the game of baseball. The Tigers wear a white and navy blue home jersey with "Tigres" across the chest for their annual "¡Fiesta Tigres!" game to recognize and honor the contributions of Hispanic and Latino players and coaches to the game of baseball. Like all of MLB, the Tigers wore a highly stylized and brightly colored jersey for Players Weekend. In the inaugural games from August 25–27, 2017, their away jerseys were grey with bright orange with "Tigers" on the chest, the orange cap had a tiger instead of the Old English "D" on it. Players were also encouraged to use nicknames on the back of their jerseys. In the 2018 season, the Tigers wore a blue jersey with orange sleeves that said "Tigers" on the chest, with an orange cap that had a tiger on it. Baseball Hall of Famers Ford C. Frick Award recipients Michigan Sports Hall of Fame Retired numbers and honorees This is how the retired numbers and honored names are displayed on the outfield walls at Comerica Park: In left field: In right field: Ty Cobb, Harry Heilmann, Heinie Manush, Hughie Jennings, and Sam Crawford are honored by their names on the wall at Comerica Park. They all played in an era where numbers were not worn on jerseys. Though their numbers are not officially retired, the names of Mickey Cochrane and George Kell are displayed at Comerica Park to honor their contributions to the Detroit Tigers organization. Cochrane wore No. 3, later retired for Alan Trammell, and Kell wore No. 21 for five years, 7 for two years, and 15 for one year. They both entered the National Baseball Hall of Fame and Museum as Detroit Tigers, and their plaques in the Hall show them wearing the Tigers' cap. Al Kaline was also a TV color commentator for the Tigers from 1975 to 2002. George Kell was also a play-by-play announcer for the Tigers from 1959 to 1963 and again from 1965 to 1996. Ernie Harwell spent 42 years (in two stints) calling Tigers games on radio and television. Jackie Robinson's No. 42 was retired throughout Major League Baseball in 1997. Almost all the players with retired numbers (and Ty Cobb) also have statues of themselves that sit behind their names, which are painted on the left-center field wall. National Avenue, which runs behind the third-base stands at the Tigers' previous home Tiger Stadium, was renamed Cochrane Street for Mickey Cochrane. Cherry Street, which runs behind the left-field stands at Tiger Stadium, was renamed Kaline Drive for Al Kaline. Team captains Bill Coughlin 1907–1908 Germany Schaefer 1907–1909 George Moriarty 1911–1916 Brad Ausmus 1999 Doug Brocail 1999 Dean Palmer 1999 Award winners Roster Minor league affiliations The Detroit Tigers farm system consists of seven minor league affiliates. Broadcasters Radio The Tigers' current flagship radio stations are Detroit sister stations WXYT (1270 AM) and WXYT-FM (97.1 FM). Dan Dickerson does play-by-play and former Tigers catcher Jim Price does color commentary. Games are syndicated throughout Michigan, Toledo and Archbold, Ohio. Television As of 2021, the Tigers' current exclusive local television rights holder is Bally Sports Detroit, which picked up the rights in 1998 taking them away from Pro-Am Sports System, owned by Post-Newsweek Stations. The Tigers renewed in 2008, over a bid from a rival regional sports channel by Dish Network and AT&T's U-verse, apparently until 2021. Through 25 games in 2017, their games have averaged a 5.57 rating, which was fifth in the major league. During the 2016 season, the Tigers averaged a 7.56 rating and 138,000 viewers on primetime TV broadcasts. The Tigers' television broadcast team consists of Matt Shepard on play-by-play and former Tigers players Kirk Gibson, Todd Jones, Cameron Maybin and Craig Monroe rotating on color commentary. Former Franchise records Facilities The team maintains a training center in the Dominican Republic. See also List of Detroit Tigers seasons Detroit Tigers all-time roster Managers and ownership of the Detroit Tigers Notes References Further reading External links Detroit Tigers at Baseball Reference Sports E-Cyclopedia Detroit Tigers Page Culture of Detroit Major League Baseball teams Baseball teams established in 1894 Grapefruit League Michigan culture Professional baseball teams in Michigan 1894 establishments in Michigan Olympia Entertainment
8733
https://en.wikipedia.org/wiki/Digital%20video
Digital video
Digital video is an electronic representation of moving visual images (video) in the form of encoded digital data. This is in contrast to analog video, which represents moving visual images in the form of analog signals. Digital video comprises a series of digital images displayed in rapid succession, usually at 24 frames per second. Digital video has many advantages such as easy copying, multicasting, sharing and storage. Digital video was first introduced commercially in 1986 with the Sony D1 format, which recorded an uncompressed standard-definition component video signal in digital form. In addition to uncompressed formats, popular compressed digital video formats today include H.264 and MPEG-4. Modern interconnect standards used for playback of digital video include HDMI, DisplayPort, Digital Visual Interface (DVI) and serial digital interface (SDI). Digital video can be copied and reproduced with no degradation in quality. In contrast, when analog sources are copied, they experience generation loss. Digital video can be stored on digital media such as Blu-ray Disc, on computer data storage, or streamed over the Internet to end users who watch content on a desktop computer screen or a digital smart TV. Today, digital video content such as TV shows and movies also include a digital audio soundtrack. History Digital video cameras The basis for digital video cameras are metal–oxide–semiconductor (MOS) image sensors. The first practical semiconductor image sensor was the charge-coupled device (CCD), invented in 1969 by Willard S. Boyle, who won a Nobel Prize for his work in physics. Following the commercialization of CCD sensors during the late 1970s to early 1980s, the entertainment industry slowly began transitioning to digital imaging and digital video over from analog video the next two decades. The CCD was followed by the CMOS active-pixel sensor (CMOS sensor), developed in the 1990s. CMOS are beneficial because of their small size, high speed, and low power usage. CMOS are most commonly found today in the digital cameras in iPhones, used as the image censor for the device. Digital video coding In the 1970s, pulse-code modulation (PCM) induced the birth of digital video coding, demanding high bitrates of 45-140 Mbps for standard-definition (SD) content. By the 1980s, the discrete cosine transform (DCT) became the standard for digital video compression. The first digital video coding standard was H.120, created by the (International Telegraph and Telephone Consultative Committee) or CCITT (now ITU-T) in 1984. H.120 was not practical due to weak performance. H.120 was based on differential pulse-code modulation (DPCM), a compression algorithm that was inefficient for video coding. During the late 1980s, a number of companies began experimenting with DCT, a much more efficient form of compression for video coding. The CCITT received 14 proposals for DCT-based video compression formats, in contrast to a single proposal based on vector quantization (VQ) compression. The H.261 standard was developed based on DCT compression, becoming first practical video coding standard. Since H.261, DCT compression has been adopted by all the major video coding standards that followed. MPEG-1, developed by the Motion Picture Experts Group (MPEG), followed in 1991, and it was designed to compress VHS-quality video. It was succeeded in 1994 by MPEG-2/H.262, which became the standard video format for DVD and SD digital television. It was followed by MPEG-4/H.263 in 1999, and then in 2003 it was followed by H.264/MPEG-4 AVC, which has become the most widely used video coding standard. Digital video production Starting in the late 1970s to the early 1980s, video production equipment that was digital in its internal workings was introduced. These included time base correctors (TBC) and digital video effects (DVE) units. They operated by taking a standard analog composite video input and digitizing it internally. This made it easier to either correct or enhance the video signal, as in the case of a TBC, or to manipulate and add effects to the video, in the case of a DVE unit. The digitized and processed video information was then converted back to standard analog video for output. Later on in the 1970s, manufacturers of professional video broadcast equipment, such as Bosch (through their Fernseh division) and Ampex developed prototype digital videotape recorders (VTR) in their research and development labs. Bosch's machine used a modified 1 inch type B videotape transport and recorded an early form of CCIR 601 digital video. Ampex's prototype digital video recorder used a modified 2-inch quadruplex videotape VTR (an Ampex AVR-3) fitted with custom digital video electronics and a special "octaplex" 8-head headwheel (regular analog 2" quad machines only used 4 heads). Like standard 2" quad, the audio on the Ampex prototype digital machine, nicknamed by its developers as "Annie," still recorded the audio in analog as linear tracks on the tape. None of these machines from these manufacturers were ever marketed commercially. Digital video was first introduced commercially in 1986 with the Sony D1 format, which recorded an uncompressed standard definition component video signal in digital form. Component video connections required 3 cables, but most television facilities were wired for composite NTSC or PAL video using one cable. Due this incompatibility the cost of the recorder, D1 was used primarily by large television networks and other component-video capable video studios. In 1988, Sony and Ampex co-developed and released the D2 digital videocassette format, which recorded video digitally without compression in ITU-601 format, much like D1. In comparison, D2 had the major difference of encoding the video in composite form to the NTSC standard, thereby only requiring single-cable composite video connections to and from a D2 VCR. This made it a perfect fit for the majority of television facilities at the time. D2 was a successful format in the television broadcast industry throughout the late '80s and the '90s. D2 was also widely used in that era as the master tape format for mastering laserdiscs. D1 & D2 would eventually be replaced by cheaper systems using video compression, most notably Sony's Digital Betacam, that were introduced into the network's television studios. Other examples of digital video formats utilizing compression were Ampex's DCT (the first to employ such when introduced in 1992), the industry-standard DV and MiniDV and its professional variations, Sony's DVCAM and Panasonic's DVCPRO, and Betacam SX, a lower-cost variant of Digital Betacam using MPEG-2 compression. One of the first digital video products to run on personal computers was PACo: The PICS Animation Compiler from The Company of Science & Art in Providence, RI. It was developed starting in 1990 and first shipped in May 1991. PACo could stream unlimited-length video with synchronized sound from a single file (with the ".CAV" file extension) on CD-ROM. Creation required a Mac, and playback was possible on Macs, PCs, and Sun SPARCstations. QuickTime, Apple Computer's multimedia framework, was released in June 1991. Audio Video Interleave from Microsoft followed in 1992. Initial consumer-level content creation tools were crude, requiring an analog video source to be digitized to a computer-readable format. While low-quality at first, consumer digital video increased rapidly in quality, first with the introduction of playback standards such as MPEG-1 and MPEG-2 (adopted for use in television transmission and DVD media), and the introduction of the DV tape format allowing recordings in the format to be transferred directly to digital video files using a FireWire port on an editing computer. This simplified the process, allowing non-linear editing systems (NLE) to be deployed cheaply and widely on desktop computers with no external playback or recording equipment needed. The widespread adoption of digital video and accompanying compression formats has reduced the bandwidth needed for a high-definition video signal (with HDV and AVCHD, as well as several commercial variants such as DVCPRO-HD, all using less bandwidth than a standard definition analog signal). These savings have increased the number of channels available on cable television and direct broadcast satellite systems, created opportunities for spectrum reallocation of terrestrial television broadcast frequencies, and made tapeless camcorders based on flash memory possible, among other innovations and efficiencies. Digital video and culture Culturally, digital video has allowed video and film to become widely available and popular, beneficial to entertainment, education, and research. Digital video is increasingly common in schools, with students and teachers taking an interest in learning how to use it in relevant ways. Digital video also has healthcare applications, allowing doctors to track infant heart rates and oxygen levels. In addition, the switch from analog to digital video impacted media in various ways, such as in how business use cameras for surveillance. Closed circuit television (CCTV) switched to using digital video recorders (DVR), presenting the issue of how to store recordings for evidence collection. Today, digital video is able to be compressed in order to save storage space. Digital television Digital Television, also known as DTV, is the production and transmission of digital video from networks to consumers. This technique uses digital encoding instead of analog signals used prior to the 1950s. As compared to analog methods, DTV is faster and provides more capabilities and options for data to be transmitted and shared. Overview Digital video comprises a series of digital images displayed in rapid succession. In the context of video, these images are called frames. The rate at which frames are displayed is known as the frame rate and is measured in frames per second (FPS). Every frame is a digital image and so comprises a formation of pixels. The color of a pixel is represented by a fixed number of bits of that color where the information of the color is stored within the image. For example, 8-bit captures 256 levels per channel, and 10-bit captures 1,024 levels per channel. The more bits, the more subtle variations of colors can be reproduced. This is called the color depth, or bit depth, of the video. Interlacing In interlaced video each frame is composed of two halves of an image. The first half contains only the odd-numbered lines of a full frame. The second half contains only the even-numbered lines. These halves are referred to individually as fields. Two consecutive fields compose a full frame. If an interlaced video has a frame rate of 30 frames per second the field rate is 60 fields per second, though both part of interlaced video, frames per second and fields per second are separate numbers. Bit rate and BPP By definition, bit rate is a measurement of the rate of information content from the digital video stream. In the case of uncompressed video, bit rate corresponds directly to the quality of the video because bit rate is proportional to every property that affects the video quality. Bit rate is an important property when transmitting video because the transmission link must be capable of supporting that bit rate. Bit rate is also important when dealing with the storage of video because, as shown above, the video size is proportional to the bit rate and the duration. Video compression is used to greatly reduce the bit rate while having little effect on quality. Bits per pixel (BPP) is a measure of the efficiency of compression. A true-color video with no compression at all may have a BPP of 24 bits/pixel. Chroma subsampling can reduce the BPP to 16 or 12 bits/pixel. Applying JPEG compression on every frame can reduce the BPP to 8 or even 1 bits/pixel. Applying video compression algorithms like MPEG1, MPEG2 or MPEG4 allows for fractional BPP values to exist. Constant bit rate versus variable bit rate BPP represents the average bits per pixel. There are compression algorithms that keep the BPP almost constant throughout the entire duration of the video. In this case, we also get video output with a constant bitrate (CBR). This CBR video is suitable for real-time, non-buffered, fixed bandwidth video streaming (e.g. in videoconferencing). Since not all frames can be compressed at the same level, because quality is more severely impacted for scenes of high complexity, some algorithms try to constantly adjust the BPP. They keep the BPP high while compressing complex scenes and low for less demanding scenes. This way, it provides the best quality at the smallest average bit rate (and the smallest file size, accordingly). This method produces a variable bitrate because it tracks the variations of the BPP. Technical overview Standard film stocks typically record at 24 frames per second. For video, there are two frame rate standards: NTSC, at 30/1.001 (about 29.97) frames per second (about 59.94 fields per second), and PAL, 25 frames per second (50 fields per second). Digital video cameras come in two different image capture formats: interlaced and progressive scan. Interlaced cameras record the image in alternating sets of lines: the odd-numbered lines are scanned, and then the even-numbered lines are scanned, then the odd-numbered lines are scanned again, and so on. One set of odd or even lines is referred to as a field, and a consecutive pairing of two fields of opposite parity is called a frame. Progressive scan cameras record all lines in each frame as a single unit. Thus, interlaced video captures the scene motion twice as often as progressive video does for the same frame rate. Progressive-scan generally produces a slightly sharper image, however, motion may not be as smooth as interlaced video. Digital video can be copied with no generation loss; which degrades quality in analog systems. However, a change in parameters like frame size, or a change of the digital format can decrease the quality of the video due to image scaling and transcoding losses. Digital video can be manipulated and edited on non-linear editing systems. Digital video has a significantly lower cost than 35 mm film. In comparison to the high cost of film stock, the digital media used for digital video recording, such as flash memory or hard disk drive is very inexpensive. Digital video also allows footage to be viewed on location without the expensive and time-consuming chemical processing required by film. Network transfer of digital video makes physical deliveries of tapes and film reels unnecessary. Digital television (including higher quality HDTV) was introduced in most developed countries in early 2000s. Today, digital video is used in modern mobile phones and video conferencing systems. Digital video is used for Internet distribution of media, including streaming video and peer-to-peer movie distribution. Many types of video compression exist for serving digital video over the internet and on optical disks. The file sizes of digital video used for professional editing are generally not practical for these purposes, and the video requires further compression with codecs to be used for recreational purposes. , the highest image resolution demonstrated for digital video generation is 132.7 megapixels (15360 x 8640 pixels). The highest speed is attained in industrial and scientific high speed cameras that are capable of filming 1024x1024 video at up to 1 million frames per second for brief periods of recording. Technical properties Live digital video consumes bandwidth. Recorded digital video consumes data storage. The amount of bandwidth or storage required is determined by the frame size, color depth and frame rate. Each pixel consumes a number of bits determined by the color depth. The data required to represent a frame of data is determined by multiplying by the number of pixels in the image. The bandwidth is determined by multiplying the storage requirement for a frame by the frame rate. The overall storage requirements for a program can then be determined by multiplying bandwidth by the duration of the program. These calculations are accurate for uncompressed video, but due to the relatively high bit rate of uncompressed video, video compression is extensively used. In the case of compressed video, each frame requires only a small percentage of the original bits. This reduces the data or bandwidth consumption by a factor of 5 to 12 times when using lossless compression, but more commonly, lossy compression is used due to its reduction of data consumption by factors of 20 to 200. Note that it is not necessary that all frames are equally compressed by the same percentage. Instead, consider the average factor of compression for all the frames taken together. Interfaces and cables Purpose-built digital video interfaces Digital component video Digital Visual Interface (DVI) DisplayPort HDBaseT High-Definition Multimedia Interface (HDMI) Unified Display Interface General-purpose interfaces use to carry digital video FireWire (IEEE 1394) Universal Serial Bus (USB) The following interface has been designed for carrying MPEG-Transport compressed video: DVB-ASI Compressed video is also carried using UDP-IP over Ethernet. Two approaches exist for this: Using RTP as a wrapper for video packets as with SMPTE 2022 1–7 MPEG Transport Packets are placed directly in the UDP packet Other methods of carrying video over IP Network Device Interface SMPTE 2110 Storage formats Encoding CCIR 601 used for broadcast stations VC-2 also known as Dirac Pro MPEG-4 good for online distribution of large videos and video recorded to flash memory MPEG-2 used for DVDs, Super-VCDs, and many broadcast television formats MPEG-1 used for video CDs H.261 H.263 H.264 also known as MPEG-4 Part 10, or as AVC, used for Blu-ray Discs and some broadcast television formats H.265 also known as MPEG-H Part 2, or as HEVC MOV used for QuickTime framework Theora used for video on Wikipedia Tapes Betacam SX, Betacam IMX, Digital Betacam, or DigiBeta — commercial video systems by Sony, based on original Betamax technology D-VHS — MPEG-2 format data recorded on a tape similar to S-VHS D1, D2, D3, D5, D9 (also known as Digital-S) — various SMPTE commercial digital video standards Digital8 — DV-format data recorded on Hi8-compatible cassettes; largely a consumer format DV, MiniDV — used in most of today's videotape-based consumer camcorders; designed for high quality and easy editing; can also record high-definition data (HDV) in MPEG-2 format DVCAM, DVCPRO — used in professional broadcast operations; similar to DV but generally considered more robust; though DV-compatible, these formats have better audio handling. DVCPRO50, DVCPROHD support higher bandwidths as compared to Panasonic's DVCPRO. HDCAM was introduced by Sony as a high-definition alternative to DigiBeta. MicroMV — MPEG-2-format data recorded on a very small, matchbook-sized cassette; obsolete ProHD — name used by JVC for its MPEG-2-based professional camcorders Discs Blu-ray Disc DVD VCD See also Digital audio Digital cinematography Display aspect ratio Display resolution Index of video-related articles Internet video Online video platform Video coding format Video editing software Webcam Notes References External links The DV, DVCAM, & DVCPRO Formats – tech details, FAQ, and links Standard digital TV and video formats. Film and video technology Video signal Television terminology Audiovisual introductions in 1986 Film and video terminology pl:Digital video pt:Vídeo digital
8892
https://en.wikipedia.org/wiki/Dionysius%20Thrax
Dionysius Thrax
Dionysius Thrax ( Dionýsios ho Thrâix, 170–90 BC) was a Greek grammarian and a pupil of Aristarchus of Samothrace. He was long considered to be the author of the earliest grammatical text on the Greek language, one that was used as a standard manual for perhaps some 1,500 years, and which was until recently regarded as the groundwork of the entire Western grammatical tradition. Life His place of origin was not Thrace as the epithet "Thrax" denotes, but probably Alexandria. His Thracian background was inferred from the name of his father Tērēs (Τήρης), which is considered to be a Thracian name. One of his co-students during his studies in Alexandria under Aristarchus was Apollodorus of Athens, who also became a distinguished grammarian. Rudolf Pfeiffer dates his shift to the isle of Rhodes to around 144/143 BC, when political upheavals associated with the policies of Ptolemy VIII Euergetes II are thought to have led to his exile. According to a report in Athenaeus' Deipnosophistae (11,489a, b), his Rhodian pupils, grateful for his learning, gathered enough silver to enable him to fashion a cup whose shape aspired to recreate that of Nestor mentioned in the Iliad (Book 11, lines 632–637). Dionysius was primarily an Homeric scholar, which was integral to his training under Aristarchus in Alexandria. His work shows some influence of earlier Stoic grammatical theory, particular on word classes. He is also reported by Varro to have been an erudite analyst of Greek lyric poetry, perhaps referring to his linguistic and prosodic use of that material. He wrote prolifically in three genres: philological questions (γραμματικά); running commentaries (ὑπομνήματα) and treatises (συνταγματικά). Of the last genre, he wrote a polemical monograph criticizing the Homeric interpretations of Krates. Another work he is said to have written was the Περὶ ποσοτήτων (On quantities). From the scholia preserved from the critical works of Aristonicus and Didymus who excerpted Dionysius' work it is clear that he was decidedly independent in his textual judgements on the Homeric corpus, since he frequently contradicts his master's known readings. His teaching may have exercised a formative impact on the rise of Roman grammatical studies, if as an entry in the Suda suggests, the elder Tyrannion was one of his pupils. The founder of classical scholarship in Rome, L. Aelius Stilo may have profited from Dionysius' instruction, since he accompanied to Rhodes Q. Metellus Numidicus when the latter went into voluntary exile, and while Dionysius was still teaching there. Tékhnē grammatikē Dionysius Thrax was credited traditionally as the author of the first extant grammar of Greek, Art of Grammar (, Tékhnē grammatikē). The Greek text, in August Immanuel Bekker's edition, runs to fifty pages. Its importance in Byzantine scholarship is attested by the fact that commentaries on it by Byzantine scholiasts run to some 600 pages. The text itself was thought to be the unique extant example of a work by Hellenistic scholars. This general consensus began to break down when examinations of grammatical texts datable to a later period emerged among the finds of the Oxyrhynchus Papyri which, until relatively late, showed no awareness of key elements in the text attributed to Dionysius Thrax. It concerns itself primarily with a morphological description of Greek, lacking any treatment of syntax. The work was translated into Armenian sometime around the 5th to 6th centuries AD, and into Syriac by Joseph Huzaya around that same period. Dionysius defines grammar at the beginning of the Tékhnē as "the empirical knowledge of what is for the most part being said by poets and prose writers". He states that grammatikē, what we might nowadays call "literary criticism", comprises six parts: Grammatikḗ (a) ἀνάγνωσις ἐντριβὴς κατὰ προσῳδίαν (anagnōsis...): reading aloud with correct pronunciation, accent and punctuation. (b) ἐξήγησις κατὰ τοὺς ἐνυπάρχοντας ποιητικοὺς τρόπους (exēgēsis...): exposition of the tropes/τρόποι, the figurative language of texts. (c) ἀπόδοσις πρόχειρος γλωσσῶν τε καὶ ἰστοριῶν (apodosis...): common exposition of obsolete words and subject matter. (d) εὕρεσις ἐτυμολογίας (heuresis...): finding the correct meaning of words according to their origin (etymology). (e) ἐκλογισμὸς ἀναλογίας (eklogismos...): setting forth or considering analogies. (f) κρίσις ποιημάτων (krisis...): critical judgement of the works examined. Paragraph 6 outlines the στοιχεῖα (stoikheia) or letters of the alphabet, together with the divisions into vowels, diphthongs and consonants. Paragraphs 7–10 deal with syllables, long (μακραὶ συλλαβαί), short (βραχεῖαι συλλαβαί) and anceps (κοιναὶ συλλαβαί). Paragraph 11 treats the eight word classes, though strong doubts exist as to whether or not this division goes back to Dionysius Thrax, since ancient testimonies assert that he conflated proper nouns and appellatives, and classified the article together with pronouns. In the text attributed to Dionysius, the eight classes. which Di Benedetto and others argue was probably developed by Tryphon several decades after Dionysius, are as follows: (a) the 'name' (ὄνομα ónoma), translated as noun: a part of speech inflected for case. Its three genders: masculine (ἀρσενικόν), feminine (θηλυκόν) and neutral (οὐδέτερον) are distinguished, together with the five case endings. He also notes however that two other terms are also in use: κοινόν (common) designating those words whose gender varies depending on the sex of the creature, such as ἵππος (hippos 'horse') and ἐπίκοινον (epicene) used to define words whose gender is stable, but which can refer to either sex, instancing χελιδών (khelidōn 'swallow'). The name includes various species like nouns, adjectives, proper nouns, appellatives, collectives, ordinals, numerals and more. For example, the appellative (προσηγορία), which he considers a species (εἶδος) of the proper noun, not a distinct part of speech. (b) the verb (ῥῆμα) with its tenses. (c) the participle (μετοχή) (d) the article (ἄρθρον) (e) the pronoun (ἀντωνυμία) (f) the preposition (πρόθεσις) (g) the adverb (ἐπίρρημα) (h) the conjunction (σύνδεσμος) Paragraphs 12-20 then elaborates successively on the parts of speech. Authorship Modern scepticism over the attribution is associated with the pioneering work of Vincenzo Di Benedetto in particular, though as early as 1822 Karl Wilhelm Göttling, by analyzing the scholia on the text that had recently been collected and published by A. I. Bekker, concluded that the text as we have it was to be dated, not to the Hellenistic period but rather to the Byzantine period. Göttling's thesis convinced neither Moritz Schmidt nor Gustav Uhlig, and disappeared from view. In 1958/1959, Di Benedetto revived doubts by comparing the received text with ancient grammatical papyri that had since come to light. He argued that before the 3rd to 4th centuries AD, no papyri on Greek grammar reveal material structured in a way similar to the exposition we have in Dionysius's treatise, that the surviving witnesses for the period before that late date, namely authors such as Sextus Empiricus, Aelius Herodianus, Apollonius Dyscolus and Quintilian, fail to cite him, and that Dionysius's work only begins to receive explicit mention in the works written from the 5th century onwards by such scholars as Timotheus of Gaza, Ammonius Hermiae and Priscian. Di Benedetto concluded that only the first five paragraphs of the treatise came from Dionysius' hand. Though initially rebuffed by scholars of the caliber of Pfeiffer and Hartmut Erbse, Di Benedetto's argument today has found general acceptance among specialists. Notes Citations Sources External links . 170 BC births 90 BC deaths Ancient Greek grammarians Ancient Thracian Greeks Koine Greek Linguists from Greece Thracian people Ancient Alexandrians
9091
https://en.wikipedia.org/wiki/Doctor%20V64
Doctor V64
The Doctor V64 (also referred to simply as the V64) is a development and backup device made by Bung Enterprises Ltd that is used in conjunction with the Nintendo 64. The Doctor V64 also had the ability to play video CDs and audio CDs. Additionally, it could apply stereo 3D effects to the audio. History The V64 was released in 1996 and was priced around $450 USD. It was the first commercially-available backup device for the Nintendo 64 unit. The Partner N64 development kit, which was manufactured by Silicon Graphics and sold officially by Nintendo, was a comparatively expensive development machine. The V64 served as a lower-cost development machine, though its unofficial status would later lead to conflict with Nintendo. Some third-party developers used a number of V64s in their development process. Specifications The CPU of the V64 is a 6502, and the operating system is contained in a BIOS. The V64 unit contains a CD-ROM drive which sits underneath the Nintendo 64 and plugs into the expansion slot on the underside of the Nintendo 64. The expansion slot is essentially a mirror image of the cartridge slot on the top of the unit, with the same electrical connections; thus, the Nintendo 64 reads data from the Doctor V64 in the same manner as it would from a cartridge plugged into the normal slot. Usage Game booting In order to get around Nintendo's lockout chip, when using the V64, a game cartridge is plugged into the Nintendo 64 through an adaptor which connects only the lockout chip. The game cart used for the operation had to contain the same lockout chip used by the game back up. Saving game progress The second problem concerned saving progress. Most N64 games are saved to the cart itself instead of external memory cards. If the player wanted to keep their progress, then the cartridge used had to have the same type of non-volatile memory hardware. Alternatively, Bung produced the "DX256" and "DS1" add-ons to allow (EEPROM and SRAM respectively) saves to be made without using the inserted cartridge. These devices were inserted into the top-slot of the N64 with the game cartridge being then inserted into the top of them to just provide the security bypass. Save slots on the DX256 were selected using an alpha and numeric encoder knobs on the front of the device. Uploading game images The Doctor V64 could be used to read the data from a game cartridge and transfer the data to a PC via the parallel port. This allowed developers and homebrew programmers to upload their game images to the Doctor V64 without having to create a CD backup each time. It also allowed users to upload game images taken from the Internet. Doctor V64 Jr. Following the Doctor V64's success, Bung released the Doctor V64 Jr. in December 1998. This was a condensed, cost-efficient version of the original V64. The Doctor V64 Jr. has no CD drive and plugs into the normal cartridge slot on the top of the Nintendo 64. Data is loaded into the Doctor V64 Jr.'s battery-backed RAM from a PC via a parallel port connection. The Doctor V64 Jr. has up to 512 megabits (64 MB) of memory storage. This was done to provide for future Nintendo 64 carts that employed larger memory storage, but the high costs associated with ordering large storage carts kept this occurrence at a minimum. Only a handful of 512-megabit games were released for the Nintendo 64 system. Promotions In 1998 and 1999, there was a homebrew competition known as "Presence of Mind" (POM), an N64 demo competition led by dextrose.com. The contest consisted of submitting a user-developed N64 program, game, or utility. Bung Enterprises promoted the event and supplied prizes (usually Doctor V64 related accessories). Though a contest was planned for 2000, the interest in the N64 was already fading, and so did the event. POM contest demo entries can still be found on the Internet. Legal issues Role in piracy The Doctor V64 unit was the first commercially available backup device for the Nintendo 64 unit. Though the unit was sold as a development machine, it could be modified to enable the creation and use of commercial game backups. Unlike official development units, the purchase of V64s was not restricted to software companies only. For this reason, the unit became a popular choice among those looking to proliferate unlicensed copies of games. Original Doctor V64 units sold by Bung did not allow the playing of backups. A person would have to modify the unit by themselves in order to make it backup friendly. This usually required a user to download and install a modified Doctor V64 BIOS. Additionally, the cartridge adapter had to be opened and soldered in order to allow for the operational procedure. Though Bung never sold backup enabled V64s, many re-sellers would modify the units themselves. Conflicts with Nintendo During the N64's lifetime, Nintendo revised the N64's model, making the serial port area smaller. This slight change in the N64's plastic casing made the connection to the Doctor V64 difficult to achieve without user modification. This revision may have been a direct reaction from Nintendo to discourage the use of V64 devices, and may also explain why Bung decided to discontinue the use of this port in the later Doctor V64 Jr. models. Nintendo made many legal efforts worldwide in order to stop the sale of Doctor V64 units. They sued Bung directly as well as specific store retailers in Europe and North America for copyright infringement. Eventually, Nintendo managed to have the courts prohibit the sale of Doctor V64 units in the United States. Main menu The Doctor V64 implemented text-based menu-driven screens. The menus consisted of white text superimposed over a black background. Utilizing the buttons on the V64 unit, a user would navigate the menus and issue commands. Though the menu was mainly designed for game developers, it is possible to back up cartridges with it (through the use of an unofficial V64 BIOS). Some of the menu items related to game backups were removed from the V64's BIOS near the end of its life due to pressure from Nintendo. These items are only available by obtaining a patched V64 BIOS. Detailed specifications CD-ROM access speed Most early V64 models shipped with a standard IDE 8X CD-ROM . During the manufacturing lifetime of the device, latter V64 models shipped with 16X and eventually 20X drives. V64 units could be purchased without a CD-ROM drive. It is possible to replace the unit with a faster IDE CD-ROM unit (such as the 52X model in the image on this page). Many Doctor V64s shipped internationally were ordered without an installed CD-ROM drive, to save on shipping costs associated with weight, to avoid import duty on the drive, and to allow users to customize the units in response to the ever-increasing speeds of drives available. The variance in the power draw of different manufacturers drives at different speeds caused issues with disc spin-ups exceeding the wattage rating of the included Bung PSU. This led to users swapping out the Bung PSU for a more powerful model, or selecting low draw drives (mainly Panasonic drives sometimes badged as Creative). CD-Media V64s can read CD-Rs and CD-RWs (provided the installed CD-ROM unit supports rewritable media). Supported media has to be recorded in Mode 1, ISO 9660 format. Doctor V64s only support the 8.3 DOS naming convention. As such, Joliet file system is not supported. RAM Depending on the model, V64s came with either 128 megabits (16 MB) or 256 megabits (32 MB) of RAM. Original V64 units shipped with 128 megabits of RAM. V64 units started shipping with 256 megabits when developers started using bigger sized memory carts for their games. Users had the option of buying a memory upgrade from Bung and other re-sellers. Power supply The Doctor V64 uses a 4 Pin MiniDIN jack (as used for S-Video) for connecting the power supply cord. Power supplies included with Doctor V64s were very unreliable. Bung replaced the power supply with a sturdier version in later V64 units. Replacing broken power supplies became one of the most common maintenance problems with the V64. It is possible to modify an AT PC power supply for V64 use. Only 4 cables have to be connected to the V64 for it to function. Additional information The ROM extensions ".v64" and ".z64" started out as the preferred naming conventions by Doctor V64 and Z64 users, respectively. It would also imply the file's "endianness" as those units employed little endian (V64) and big endian (Z64) byte alignment. ".n64" was used as well but not as much (it became more popular as N64 emulators began to appear). The terms ".v64" and ".z64" are still widely used today by the emulation community. Acclaim Entertainment subsidiary Iguana Entertainment used Doctor V64 units as their development hardware of choice during the N64 era. They were best known for developing the Turok, NBA Jam, NFL Quarterback Club, and South Park video games. References Nintendo 64 accessories Unlicensed Nintendo hardware
9317
https://en.wikipedia.org/wiki/European%20Union
European Union
The European Union (EU) is a supranational political and economic union of member states that are located primarily in Europe. The union has a total area of and an estimated total population of nearly 447million. The EU has often been described as a sui generis political entity (without precedent or comparison) combining the characteristics of both a federation and a confederation. Containing 5.8 per cent of the world population in 2020, the EU generated a nominal gross domestic product (GDP) of around trillion in 2022, constituting approximately one sixth of global nominal GDP and the third-biggest global economy after the United States and China. Additionally, all EU states except Bulgaria have a very high Human Development Index according to the United Nations Development Programme. Its cornerstone, the Customs Union, paved the way to establishing an internal single market based on standardised legal framework and legislation that applies in all member states in those matters, and only those matters, where the states have agreed to act as one. EU policies aim to ensure the free movement of people, goods, services and capital within the internal market; enact legislation in justice and home affairs; and maintain common policies on trade, agriculture, fisheries and regional development. Passport controls have been abolished for travel within the Schengen Area. The eurozone is a group composed of the 20 EU member states that have fully implemented the economic and monetary union and use the euro currency. Through the Common Foreign and Security Policy, the union has developed a role in external relations and defence. It maintains permanent diplomatic missions throughout the world and represents itself at the United Nations, the World Trade Organization, the G7 and the G20. Due to its global influence, the European Union has been described by some scholars as an emerging superpower. The union was established along with its citizenship when the Maastricht Treaty came into force in 1993, and was subsequently incorporated as an international law juridical person upon entry into force of the Treaty of Lisbon in 2009, but its beginnings may be traced to its earliest predecessors incorporated primarily by a group of founding states known as the Inner Six (Belgium, France, Italy, Luxembourg, the Netherlands, and West Germany) at the start of modern institutionalised European integration in 1948 and onwards, namely to the Western Union (WU, 1954 renamed Western European Union, WEU), the International Authority for the Ruhr (IAR), the European Coal and Steel Community (ECSC), the European Economic Community (EEC, 1993 renamed European Community, EC) and the European Atomic Energy Community (Euratom), established, respectively, by the 1948 Treaty of Brussels, the 1948 London Six-Power Conference, the 1951 Treaty of Paris, the 1957 Treaty of Rome and the 1957 Euratom Treaty. These increasingly amalgamated bodies, later known collectively as the European Communities have grown since, along with their legal successor, the EU, both in size through accessions of further 21 states as well as in power through acquisitions of various policy areas to their remit by the virtue of the abovementioned treaties, as well as numerous other ones, such as the Modified Brussels Treaty, the Merger Treaty, the Single European Act, the Treaty of Amsterdam and the Treaty of Nice. In 2012, the EU was awarded the Nobel Peace Prize. After the creation by six states, 22 other states joined the union in 1973–2013. The United Kingdom became the only member state to leave the EU in 2020; ten countries are aspiring or negotiating to join it. History Origins After the First World War, but particularly with Second World War, internationalism was gaining, with the creation of the Bretton Woods System in 1944, the United Nations in 1945 and the French Union (19461958), the latter directing decolonization by possibly integrating its colonies into a European community. In this light European integration was seen, already during the war, as an antidote to the extreme nationalism which had devastated parts of the continent. The Ventotene prison Manifesto of 1941 by Altiero Spinelli propagated European integration through the Italian Resistance and after 1943 through the European Federalist Movement. Winston Churchill called in 1943 for a post-war "Council of Europe" and on 19 September 1946, at the University of Zürich, coincidentally parallel to the Hertenstein Congress of the Union of European Federalists, for a United States of Europe. Richard von Coudenhove-Kalergi, who successfully established during the interwar period the oldest organization for European integration, the Paneuropean Union, founded in June 1947 the European Parliamentary Union (EPU). Towards the end of World War II, the Three Allied Powers discussed during the Tehran Conference and the ensuing 1943 Moscow Conference the plans to establish joint institutions. This led to a decision at the Yalta Conference in 1944 to include Free France as the Fourth Allied Power and to form a European Advisory Commission, later replaced by the Council of Foreign Ministers and the Allied Control Council, following the German surrender and the Potsdam Agreement in 1945. The growing rift among the Four Powers became evident as a result of the rigged 1947 Polish legislative election which constituted an open breach of the Yalta Agreement, followed by the announcement of the Truman Doctrine on 12 March 1947. On 4 March 1947, France and the United Kingdom signed the Treaty of Dunkirk for mutual assistance in the event of future military aggression in the aftermath of World War II against any of the pair. The rationale for the treaty was the threat of a potential future military attack, specifically a Soviet one in practice, though publicised under the disguise of a German one, according to the official statements. Immediately following the February 1948 coup d'état by the Communist Party of Czechoslovakia, the London Six-Power Conference was held, resulting in the Soviet boycott of the Allied Control Council and its incapacitation, an event marking the beginning of the Cold War. The remainder of the year 1948 marked the beginning of institutionalised modern European integration. Initial years and the Paris Treaty (19481957) The year 1948 marked the beginning of the institutionalised modern European integration. In March 1948 the Treaty of Brussels was signed, establishing the Western Union (WU), followed by the International Authority for the Ruhr. Furthermore, the Organisation for European Economic Co-operation (OEEC), the predecessor of the OECD, was also founded in 1948 to manage the Marshall Plan, triggering as a Soviet response formation of the Comecon. The ensuing Hague Congress of May 1948 was a pivotal moment in European integration, as it led to the creation of the European Movement International, the College of Europe and most importantly to the foundation of the Council of Europe on 5 May 1949 (today its Europe day). The Council of Europe was one of the first institutions to bring the sovereign nations of (then only Western) Europe together, raising great hopes and fevered debates in the following two years for further European integration. It has since been a broad forum to further cooperation and shared issues, achieving for example the European Convention on Human Rights in 1950. Essential for the actual birth of the institutions of the EU was the Schuman Declaration on 9 May 1950 (the day after the fifth Victory in Europe Day) and the decision by six nations (France, Belgium, Netherlands, Luxembourg, West Germany and Italy) to follow Schuman and draft the Treaty of Paris. This treaty was created in 1952 the European Coal and Steel Community (ECSC), which was built on the International Authority for the Ruhr, installed by the Western Allies in 1949 to regulate the coal and steel industries of the Ruhr area in West Germany. Backed by the Marshall Plan with large funds coming from the United States since 1948, the ECSC became a milestone organization, enabling European economic development and integration and being the origin of the main institutions of the EU such as the European Commission and Parliament. Founding fathers of the European Union understood that coal and steel were the two industries essential for waging war, and believed that by tying their national industries together, a future war between their nations became much less likely. In parallel with Schuman, the Pleven Plan of 1951 tried but failed to tie the institutions of the developing European community under the European Political Community, which was to include the also proposed European Defence Community, an alternative to West Germany joining NATO which was established in 1949 under the Truman Doctrine. In 1954 the Modified Brussels Treaty transformed the Western Union into the Western European Union (WEU). West Germany eventually joined 1955 both WEU and NATO, prompting the Soviet Union to form the Warsaw Pact in 1955 as an institutional framework for its military domination in the countries of Central and Eastern Europe. Assessing the progress of European integration the Messina Conference was held in 1955, ordering the Spaak report, which in 1956 recommended the next significant steps of European integration. Treaty of Rome (19581972) In 1957, Belgium, France, Italy, Luxembourg, the Netherlands, and West Germany signed the Treaty of Rome, which created the European Economic Community (EEC) and established a customs union. They also signed another pact creating the European Atomic Energy Community (Euratom) for cooperation in developing nuclear power. Both treaties came into force in 1958. Although the EEC and Euratom were created separately from the ECSC, they shared the same courts and the Common Assembly. The EEC was headed by Walter Hallstein (Hallstein Commission) and Euratom was headed by Louis Armand (Armand Commission) and then Étienne Hirsch (Hirsch Commission). The OEEC was in turn reformed in 1961 into the Organisation for Economic Co-operation and Development (OECD) and its membership was extended to non-European states, the US and Canada. During the 1960s, tensions began to show, with France seeking to limit supranational power. Nevertheless, in 1965 an agreement was reached, and on 1 July 1967 the Merger Treaty created a single set of institutions for the three communities, which were collectively referred to as the European Communities. Jean Rey presided over the first merged commission (Rey Commission). First enlargement, and European co-operation (19731993) In 1973, the communities were enlarged to include Denmark (including Greenland), Ireland, and the United Kingdom. Norway had negotiated to join at the same time, but Norwegian voters rejected membership in a referendum. The Ostpolitik and the ensuing détente led to establishment of a first truly pan-European body, the Conference on Security and Co-operation in Europe (CSCE), predecessor of the modern Organization for Security and Co-operation in Europe (OSCE). In 1979, the first direct elections to the European Parliament were held. Greece joined in 1981. In 1985, Greenland left the Communities, following a dispute over fishing rights. During the same year, the Schengen Agreement paved the way for the creation of open borders without passport controls between most member states and some non-member states. In 1986, the European flag began to be used by the EEC and the Single European Act was signed. Portugal and Spain joined in 1986. In 1990, after the fall of the Eastern Bloc, the former East Germany became part of the communities as part of a reunified Germany. Treaties of Maastricht, Amsterdam and Nice (19932004) The European Union was formally established when the Maastricht Treaty—whose main architects were Horst Köhler, Helmut Kohl and François Mitterrand—came into force on 1 November 1993. The treaty also gave the name European Community to the EEC, even if it was referred to as such before the treaty. With further enlargement planned to include the former communist states of Central and Eastern Europe, as well as Cyprus and Malta, the Copenhagen criteria for candidate members to join the EU were agreed upon in June 1993. The expansion of the EU introduced a new level of complexity and discord. In 1995, Austria, Finland, and Sweden joined the EU. In 2002, euro banknotes and coins replaced national currencies in 12 of the member states. Since then, the eurozone has increased to encompass 20 countries. The euro currency became the second-largest reserve currency in the world. In 2004, the EU saw its biggest enlargement to date when Cyprus, the Czech Republic, Estonia, Hungary, Latvia, Lithuania, Malta, Poland, Slovakia, and Slovenia joined the union. Treaty of Lisbon, and Brexit (2004present) In 2007, Bulgaria and Romania became EU members. Later that year, Slovenia adopted the euro, followed by Cyprus and Malta in 2008, Slovakia in 2009, Estonia in 2011, Latvia in 2014, and Lithuania in 2015. On 1 December 2009, the Lisbon Treaty entered into force and reformed many aspects of the EU. In particular, it changed the legal structure of the European Union, merging the EU three pillars system into a single legal entity provisioned with a legal personality, created a permanent president of the European Council, the first of which was Herman Van Rompuy, and strengthened the position of the high representative of the union for foreign affairs and security policy. In 2012, the EU received the Nobel Peace Prize for having "contributed to the advancement of peace and reconciliation, democracy, and human rights in Europe". In 2013, Croatia became the 28th EU member. From the beginning of the 2010s, the cohesion of the European Union has been tested by several issues, including a debt crisis in some of the Eurozone countries, increasing migration from Africa and Asia, and the United Kingdom's withdrawal from the EU. A referendum in the UK on its membership of the European Union was held in 2016, with 51.9 per cent of participants voting to leave. The UK formally notified the European Council of its decision to leave on 29 March 2017, initiating the formal withdrawal procedure for leaving the EU; following extensions to the process, the UK left the European Union on 31 January 2020, though most areas of EU law continued to apply to the UK for a transition period which lasted until 31 December 2020. Timeline Politics The European Union operates through a hybrid system of supranational and intergovernmental decision-making, and according to the principle of conferral (which says that it should act only within the limits of the competences conferred on it by the treaties) and of subsidiarity (which says that it should act only where an objective cannot be sufficiently achieved by the member states acting alone). Laws made by the EU institutions are passed in a variety of forms. Generally speaking, they can be classified into two groups: those which come into force without the necessity for national implementation measures (regulations) and those which specifically require national implementation measures (directives). EU policy is in general promulgated by EU directives, which are then implemented in the domestic legislation of its member states, and EU regulations, which are immediately enforceable in all member states. Lobbying at the EU level by special interest groups is regulated to try to balance the aspirations of private initiatives with public interest decision-making process. Budget The European Union had an agreed budget of  billion for the year 2007 and  billion for the period 2007–2013, representing 1.10 per cent and 1.05 per cent of the EU-27's GNI forecast for the respective periods. In 1960, the budget of the European Community was 0.03 per cent of GDP. In the 2010 budget of  billion, the largest single expenditure item was "cohesion & competitiveness" with around 45 per cent of the total budget. Next was "agriculture" with approximately 31 per cent of the total. "Rural development, environment and fisheries" takes up around 11 per cent. "Administration" accounts for around 6 per cent. The "EU as a global partner" and "citizenship, freedom, security and justice" had approximately 6 per cent and 1 per cent respectively. In November 2020, two members of the union, Hungary and Poland, blocked approval to the EU's budget at a meeting in the Committee of Permanent Representatives (Coreper), citing a proposal that linked funding with adherence to the rule of law. The budget included a COVID-19 recovery fund of billion. The budget may still be approved if Hungary and Poland withdraw their vetoes after further negotiations in the council and the European Council. Bodies combatting fraud have also been established, including the European Anti-fraud Office and the European Public Prosecutor's Office. The latter is a decentralized independent body of the European Union (EU), established under the Treaty of Lisbon between 22 of the 27 states of the EU following the method of enhanced cooperation. The European Public Prosecutor's Office investigate and prosecute fraud against the budget of the European Union and other crimes against the EU's financial interests including fraud concerning EU funds of over €10,000 and cross-border VAT fraud cases involving damages above €10 million. Governance Member states retain in principle all powers except those that they have agreed collectively to delegate to the Union as a whole, though the exact delimitation has on occasions become a subject of scholarly or legal disputes. In certain fields, members have awarded exclusive competence and exclusive mandate to the Union. These are areas in which member states have entirely renounced their own capacity to enact legislation. In other areas, the EU and its member states share the competence to legislate. While both can legislate, the member states can only legislate to the extent to which the EU has not. In other policy areas, the EU can only co-ordinate, support and supplement member state action but cannot enact legislation with the aim of harmonising national laws. That a particular policy area falls into a certain category of competence is not necessarily indicative of what legislative procedure is used for enacting legislation within that policy area. Different legislative procedures are used within the same category of competence, and even with the same policy area. The distribution of competences in various policy areas between member states and the union is divided into the following three categories: The European Union has seven principal decision-making bodies, its institutions: the European Parliament, the European Council, the Council of the European Union, the European Commission, the Court of Justice of the European Union, the European Central Bank and the European Court of Auditors. Competence in scrutinising and amending legislation is shared between the Council of the European Union and the European Parliament, while executive tasks are performed by the European Commission and in a limited capacity by the European Council (not to be confused with the aforementioned Council of the European Union). The monetary policy of the eurozone is determined by the European Central Bank. The interpretation and the application of EU law and the treaties are ensured by the Court of Justice of the European Union. The EU budget is scrutinised by the European Court of Auditors. There are also a number of ancillary bodies which advise the EU or operate in a specific area. Branches of power Executive branch The European Union executive branch is organized as a directorial system, where the executive power is jointly exercised by several people. The executive branch consists of the European Council and European Commission. The European Council sets the broad political direction to the EU. It convenes at least four times a year and comprises the president of the European Council (presently Charles Michel), the president of the European Commission and one representative per member state (either its head of state or head of government). The high representative of the union for foreign affairs and security policy (presently Josep Borrell) also takes part in its meetings. Described by some as the union's "supreme political leadership", it is actively involved in the negotiation of treaty changes and defines the EU's policy agenda and strategies. Its leadership role involves solving disputes between member states and the institutions, and to resolving any political crises or disagreements over controversial issues and policies. It acts as a "collective head of state" and ratifies important documents (for example, international agreements and treaties). Tasks for the president of the European Council are ensuring the external representation of the EU, driving consensus and resolving divergences among member states, both during meetings of the European Council and over the periods between them. The European Council should not be mistaken for the Council of Europe, an international organisation independent of the EU and based in Strasbourg. The European Commission acts both as the EU's executive arm, responsible for the day-to-day running of the EU, and also the legislative initiator, with the sole power to propose laws for debate. The commission is 'guardian of the Treaties' and is responsible for their efficient operation and policing. It has 27 European commissioners for different areas of policy, one from each member state, though commissioners are bound to represent the interests of the EU as a whole rather than their home state. The leader of the 27 is the president of the European Commission (presently Ursula von der Leyen for 20192024), proposed by the European Council, following and taking into account the result of the European elections, and is then elected by the European Parliament. The President retains, as the leader responsible for the entire cabinet, the final say in accepting or rejecting a candidate submitted for a given portfolio by a member state, and oversees the commission's permanent civil service. After the President, the most prominent commissioner is the high representative of the union for foreign affairs and security policy, who is ex-officio a vice-president of the European Commission and is also chosen by the European Council. The other 26 commissioners are subsequently appointed by the Council of the European Union in agreement with the nominated president. The 27 commissioners as a single body are subject to approval (or otherwise) by vote of the European Parliament. All commissioners are first nominated by the government of the respective member state. Legislative branch The Council of the European Union (also called the Council and the "Council of Ministers", its former title) forms one half of the EU's legislature. It consists of a representative from each member state's government and meets in different compositions depending on the policy area being addressed. Notwithstanding its different configurations, it is considered to be one single body. In addition to the legislative functions, members of the council also have executive responsibilities, such as the development of a Common Foreign and Security Policy and the coordination of broad economic policies within the Union. The Presidency of the council rotates between member states, with each holding it for six months. Beginning on 1 July 2022, the position is held by the Czech Republic. The European Parliament is one of three legislative institutions of the EU, which together with the Council of the European Union is tasked with amending and approving the European Commission's proposals. 705 members of the European Parliament (MEPs) are directly elected by EU citizens every five years on the basis of proportional representation. MEPs are elected on a national basis and they sit according to political groups rather than their nationality. Each country has a set number of seats and is divided into sub-national constituencies where this does not affect the proportional nature of the voting system. In the ordinary legislative procedure, the European Commission proposes legislation, which requires the joint approval of the European Parliament and the Council of the European Union to pass. This process applies to nearly all areas, including the EU budget. The parliament is the final body to approve or reject the proposed membership of the commission, and can attempt motions of censure on the commission by appeal to the Court of Justice. The president of the European Parliament carries out the role of speaker in Parliament and represents it externally. The president and vice-presidents are elected by MEPs every two and a half years. Judicial branch The judicial branch of the European Union is formally called the Court of Justice of the European Union and consists of two courts: the Court of Justice and the General Court. The Court of Justice is the supreme court of the European Union in matters of European Union law. As a part of the Court of Justice of the European Union, it is tasked with interpreting EU law and ensuring its uniform application across all EU member states under Article 263 of the Treaty of the Functioning of the European Union (TFEU). The Court was established in 1952, and is based in Luxembourg. It is composed of one judge per member state – currently 27 – although it normally hears cases in panels of three, five or fifteen judges. The Court has been led by president Koen Lenaerts since 2015. The ECJ is the highest court of the European Union in matters of Union law, but not national law. It is not possible to appeal against the decisions of national courts in the ECJ, but rather national courts refer questions of EU law to the ECJ. However, it is ultimately for the national court to apply the resulting interpretation to the facts of any given case. Although, only courts of final appeal are bound to refer a question of EU law when one is addressed. The treaties give the ECJ the power for consistent application of EU law across the EU as a whole. The court also acts as an administrative and constitutional court between the other EU institutions and the Member States and can annul or invalidate unlawful acts of EU institutions, bodies, offices and agencies. The General Court is a constituent court of the European Union. It hears actions taken against the institutions of the European Union by individuals and member states, although certain matters are reserved for the Court of Justice. Decisions of the General Court can be appealed to the Court of Justice, but only on a point of law. Prior to the coming into force of the Lisbon Treaty on 1 December 2009, it was known as the Court of First Instance. Additional branches The European Central Bank (ECB) is one of the institutions of the monetary branch of the European Union, the prime component of the Eurosystem and the European System of Central Banks. It is one of the world's most important central banks. The ECB Governing Council makes monetary policy for the Eurozone and the European Union, administers the foreign exchange reserves of EU member states, engages in foreign exchange operations, and defines the intermediate monetary objectives and key interest rate of the EU. The ECB Executive Board enforces the policies and decisions of the Governing Council, and may direct the national central banks when doing so. The ECB has the exclusive right to authorise the issuance of euro banknotes. Member states can issue euro coins, but the volume must be approved by the ECB beforehand. The bank also operates the TARGET2 payments system. The European System of Central Banks (ESCB) consists of the ECB and the national central banks (NCBs) of all 27 member states of the European Union. The ESCB is not the monetary authority of the eurozone, because not all EU member states have joined the euro. The ESCB's objective is price stability throughout the European Union. Secondarily, the ESCB's goal is to improve monetary and financial cooperation between the Eurosystem and member states outside the eurozone. The European Court of Auditors (ECA) is the auditory branch of the European Union. It was established in 1975 in Luxembourg in order to improve EU financial management. It has 27 members (1 from each EU member-state) supported by approximately 800 civil servants. The European Personnel Selection Office (EPSO) is the civil service branch of the European Union, and is responsible for selecting staff to work for the institutions and agencies of the European Union including the European Parliament, the European Council, the Council of the European Union, the European Commission, the European Court of Justice, the Court of Auditors, the European External Action Service, the Economic and Social Committee, the Committee of the Regions and the European Ombudsman. Each institution is then able to recruit staff from among the pool of candidates selected by EPSO. On average, EPSO receives around 60,000-70,000 applications a year with around 1,500-2,000 candidates recruited by the European Union institutions. The European Ombudsman is the ombudsman branch of the European Union that holds the institutions, bodies and agencies of the EU to account, and promotes good administration. The Ombudsman helps people, businesses and organisations facing problems with the EU administration by investigating complaints, as well as by proactively looking into broader systemic issues. The current Ombudsman is Emily O'Reilly. The European Public Prosecutor's Office (EPPO) is the prosecutory branch of the European Union with juridical personality, established under the Treaty of Lisbon between 22 of the 27 states of the EU following the method of enhanced cooperation. It is based in Kirchberg, Luxembourg City alongside the Court of Justice of the European Union and the European Court of Auditors. Law Constitutionally, the EU bears some resemblance to both a confederation and a federation, but has not formally defined itself as either. (It does not have a formal constitution: its status is defined by the Treaty of European Union and the Treaty on the Functioning of the European Union). It is more integrated than a traditional confederation of states because the general level of government widely employs qualified majority voting in some decision-making among the member states, rather than relying exclusively on unanimity. It is less integrated than a federal state because it is not a state in its own right: sovereignty continues to flow 'from the bottom up', from the several peoples of the separate member states, rather than from a single undifferentiated whole. This is reflected in the fact that the member states remain the 'masters of the Treaties', retaining control over the allocation of competences to the union through constitutional change (thus retaining so-called Kompetenz-kompetenz); in that they retain control of the use of armed force; they retain control of taxation; and in that they retain a right of unilateral withdrawal under Article 50 of the Treaty on European Union. In addition, the principle of subsidiarity requires that only those matters that need to be determined collectively are so determined. Under the principle of supremacy, national courts are required to enforce the treaties that their member states have ratified, even if doing so requires them to ignore conflicting national law, and (within limits) even constitutional provisions. The direct effect and supremacy doctrines were not explicitly set out in the European Treaties but were developed by the Court of Justice itself over the 1960s, apparently under the influence of its then most influential judge, Frenchman Robert Lecourt. The question whether the secondary law enacted by the EU has a comparable status in relation to national legistaltion, has been a matter of debate among legal scholars. Primary law The European Union is based on a series of treaties. These first established the European Community and the EU, and then made amendments to those founding treaties. These are power-giving treaties which set broad policy goals and establish institutions with the necessary legal powers to implement those goals. These legal powers include the ability to enact legislation which can directly affect all member states and their inhabitants. The EU has legal personality, with the right to sign agreements and international treaties. Secondary law The main legal acts of the European Union come in three forms: regulations, directives, and decisions. Regulations become law in all member states the moment they come into force, without the requirement for any implementing measures, and automatically override conflicting domestic provisions. Directives require member states to achieve a certain result while leaving them discretion as to how to achieve the result. The details of how they are to be implemented are left to member states. When the time limit for implementing directives passes, they may, under certain conditions, have direct effect in national law against member states. Decisions offer an alternative to the two above modes of legislation. They are legal acts which only apply to specified individuals, companies or a particular member state. They are most often used in competition law, or on rulings on State Aid, but are also frequently used for procedural or administrative matters within the institutions. Regulations, directives, and decisions are of equal legal value and apply without any formal hierarchy. Foreign relations Foreign policy co-operation between member states dates from the establishment of the community in 1957, when member states negotiated as a bloc in international trade negotiations under the EU's common commercial policy. Steps for more wide-ranging co-ordination in foreign relations began in 1970 with the establishment of European Political Cooperation which created an informal consultation process between member states with the aim of forming common foreign policies. In 1987 the European Political Cooperation was introduced on a formal basis by the Single European Act. EPC was renamed as the Common Foreign and Security Policy (CFSP) by the Maastricht Treaty. The aims of the CFSP are to promote both the EU's own interests and those of the international community as a whole, including the furtherance of international co-operation, respect for human rights, democracy, and the rule of law. The CFSP requires unanimity among the member states on the appropriate policy to follow on any particular issue. The unanimity and difficult issues treated under the CFSP sometimes lead to disagreements, such as those which occurred over the war in Iraq. The coordinator and representative of the CFSP within the EU is the high representative of the union for foreign affairs and security policy who speaks on behalf of the EU in foreign policy and defence matters, and has the task of articulating the positions expressed by the member states on these fields of policy into a common alignment. The high representative heads up the European External Action Service (EEAS), a unique EU department that has been officially implemented and operational since 1 December 2010 on the occasion of the first anniversary of the entry into force of the Treaty of Lisbon. The EEAS will serve as a foreign ministry and diplomatic corps for the European Union. Besides the emerging international policy of the European Union, the international influence of the EU is also felt through enlargement. The perceived benefits of becoming a member of the EU act as an incentive for both political and economic reform in states wishing to fulfil the EU's accession criteria, and are considered an important factor contributing to the reform of European formerly Communist countries. This influence on the internal affairs of other countries is generally referred to as "soft power", as opposed to military "hard power". Humanitarian aid The European Commission's Humanitarian Aid and Civil Protection department, or "ECHO", provides humanitarian aid from the EU to developing countries. In 2012, its budget amounted to million, 51 per cent of the budget went to Africa and 20 per cent to Asia, Latin America, the Caribbean and Pacific, and 20 per cent to the Middle East and Mediterranean. Humanitarian aid is financed directly by the budget (70 per cent) as part of the financial instruments for external action and also by the European Development Fund (30 per cent). The EU's external action financing is divided into 'geographic' instruments and 'thematic' instruments. The 'geographic' instruments provide aid through the Development Cooperation Instrument (DCI, billion, 2007–2013), which must spend 95 per cent of its budget on official development assistance (ODA), and from the European Neighbourhood and Partnership Instrument (ENPI), which contains some relevant programmes. The European Development Fund (EDF, billion for the period 2008–2013 and billion for the period 2014–2020) is made up of voluntary contributions by member states, but there is pressure to merge the EDF into the budget-financed instruments to encourage increased contributions to match the 0.7 per cent target and allow the European Parliament greater oversight. In 2016, the average among EU countries was 0.4 per cent and five had met or exceeded the 0.7 per cent target: Denmark, Germany, Luxembourg, Sweden and the United Kingdom. If considered collectively, EU member states are the largest contributor of foreign aid in the world. International cooperation and development partnerships The European Union uses foreign relations instruments like the European Neighbourhood Policy which seeks to tie those countries to the east and south of the European territory of the EU to the union. These countries, primarily developing countries, include some who seek to one day become either a member state of the European Union, or more closely integrated with the European Union. The EU offers financial assistance to countries within the European Neighbourhood, so long as they meet the strict conditions of government reform, economic reform and other issues surrounding positive transformation. This process is normally underpinned by an Action Plan, as agreed by both Brussels and the target country. There is also the worldwide European Union Global Strategy. International recognition of sustainable development as a key element is growing steadily. Its role was recognised in three major UN summits on sustainable development: the 1992 UN Conference on Environment and Development (UNCED) in Rio de Janeiro, Brazil; the 2002 World Summit on Sustainable Development (WSSD) in Johannesburg, South Africa; and the 2012 UN Conference on Sustainable Development (UNCSD) in Rio de Janeiro. Other key global agreements are the Paris Agreement and the 2030 Agenda for Sustainable Development (United Nations, 2015). The SDGs recognise that all countries must stimulate action in the following key areas – people, planet, prosperity, peace and partnership – in order to tackle the global challenges that are crucial for the survival of humanity. EU development action is based on the European Consensus on Development, which was endorsed on 20 December 2005 by EU Member States, the council, the European Parliament and the commission. It is applied from the principles of Capability approach and Rights-based approach to development. Funding is provided by the Instrument for Pre-Accession Assistance and the Global Europe programmes. Partnership and cooperation agreements are bilateral agreements with non-member nations. Defence The predecessors of the European Union were not devised as a military alliance because NATO was largely seen as appropriate and sufficient for defence purposes. 22 EU members are members of NATO and Sweden is in the process of accession while the remaining member states follow policies of neutrality. The Western European Union, a military alliance with a mutual defence clause, was disbanded in 2010 as its role had been transferred to the EU. Following the Kosovo War in 1999, the European Council agreed that "the Union must have the capacity for autonomous action, backed by credible military forces, the means to decide to use them, and the readiness to do so, in order to respond to international crises without prejudice to actions by NATO". To that end, a number of efforts were made to increase the EU's military capability, notably the Helsinki Headline Goal process. After much discussion, the most concrete result was the EU Battlegroups initiative, each of which is planned to be able to deploy quickly about 1500 personnel. Since the withdrawal of the United Kingdom, France is the only member officially recognised as a nuclear weapon state and the sole holder of a permanent seat on the United Nations Security Council. France and Italy are also the only EU countries that have power projection capabilities outside of Europe. Italy, Germany, the Netherlands and Belgium participate in NATO nuclear sharing. Most EU member states opposed the Nuclear Weapon Ban Treaty. EU forces have been deployed on peacekeeping missions from middle and northern Africa to the western Balkans and western Asia. EU military operations are supported by a number of bodies, including the European Defence Agency, European Union Satellite Centre and the European Union Military Staff. The European Union Military Staff is the highest military institution of the European Union, established within the framework of the European Council, and follows on from the decisions of the Helsinki European Council (10–11 December 1999), which called for the establishment of permanent political-military institutions. The European Union Military Staff is under the authority of the High Representative of the Union for Foreign Affairs and Security Policy and the Political and Security Committee. It directs all military activities in the EU context, including planning and conducting military missions and operations in the framework of the Common Security and Defence Policy and the development of military capabilities, and provides the Political and Security Committee with military advice and recommendations on military issues. In an EU consisting of 27 members, substantial security and defence co-operation is increasingly relying on collaboration among all member states. The European Border and Coast Guard Agency (Frontex) is an agency of the EU aiming to detect and stop illegal immigration, human trafficking and terrorist infiltration. The EU also operates the European Travel Information and Authorisation System, the Entry/Exit System, the Schengen Information System, the Visa Information System and the Common European Asylum System which provide common databases for police and immigration authorities. The impetus for the development of this co-operation was the advent of open borders in the Schengen Area and the associated cross-border crime. Member states Through successive enlargements, the European Union has grown from the six founding states (Belgium, France, West Germany, Italy, Luxembourg, and the Netherlands) to members. Countries accede to the union by becoming a party to the founding treaties, thereby subjecting themselves to the privileges and obligations of EU membership. This entails a partial delegation of sovereignty to the institutions in return for representation within those institutions, a practice often referred to as "pooling of sovereignty". In some policies, there are several member states that ally with strategic partners within the union. Examples of such alliances include the Baltic Assembly, the Benelux Union, the Bucharest Nine, the Craiova Group, the EU Med Group, the Lublin Triangle, the New Hanseatic League, the Three Seas Initiative, the Visegrád Group, and the Weimar Triangle. To become a member, a country must meet the Copenhagen criteria, defined at the 1993 meeting of the European Council in Copenhagen. These require a stable democracy that respects human rights and the rule of law; a functioning market economy; and the acceptance of the obligations of membership, including EU law. Evaluation of a country's fulfilment of the criteria is the responsibility of the European Council. The four countries forming the European Free Trade Association (EFTA) are not EU members, but have partly committed to the EU's economy and regulations: Iceland, Liechtenstein and Norway, which are a part of the single market through the European Economic Area, and Switzerland, which has similar ties through bilateral treaties. The relationships of the European microstates, Andorra, Monaco, San Marino, and Vatican City include the use of the euro and other areas of co-operation. Subdivisions Subdivisions of member-states are based on the Nomenclature of Territorial Units for Statistics (NUTS), a geocode standard for statistical purposes. The standard, adopted in 2003, is developed and regulated by the European Union, and thus only covers the member states of the EU in detail. The Nomenclature of Territorial Units for Statistics is instrumental in the European Union's Structural Funds and Cohesion Fund delivery mechanisms and for locating the area where goods and services subject to European public procurement legislation are to be delivered. Schengen Area The Schengen Area is an area comprising 27 European countries that have officially abolished all passport and all other types of border control at their mutual borders. Being an element within the wider area of freedom, security and justice policy of the EU, it mostly functions as a single jurisdiction under a common visa policy for international travel purposes. The area is named after the 1985 Schengen Agreement and the 1990 Schengen Convention, both signed in Schengen, Luxembourg. Of the 27 EU member states, 23 participate in the Schengen Area. Of the four EU members that are not part of the Schengen Area, three—Bulgaria, Cyprus, and Romania—are legally obligated to join the area in the future; Ireland maintains an opt-out, and instead operates its own visa policy. The four European Free Trade Association (EFTA) member states, Iceland, Liechtenstein, Norway, and Switzerland, are not members of the EU, but have signed agreements in association with the Schengen Agreement. Also, three European microstates—Monaco, San Marino, and the Vatican City—maintain open borders for passenger traffic with their neighbours, and are therefore considered de facto members of the Schengen Area due to the practical impossibility of travelling to or from them without transiting through at least one Schengen member country. Candidate countries There are eight countries that are recognised as candidates for membership: Albania, Bosnia and Herzegovina, Moldova, Montenegro, North Macedonia, Serbia, Turkey and Ukraine. Norway, Switzerland and Iceland have submitted membership applications in the past, but subsequently frozen or withdrawn them. Additionally, Georgia and Kosovo are officially recognised as potential candidates, and have submitted membership applications. Former members Article 50 of the Lisbon Treaty provides the basis for a member to leave the EU. Two territories have left the union: Greenland (an autonomous province of Denmark) withdrew in 1985; the United Kingdom formally invoked Article 50 of the Consolidated Treaty on European Union in 2017, and became the only sovereign state to leave when it withdrew from the EU in 2020. Geography The EU's member states cover an area of , and therefore a large part of the European continent. The EU's highest peak is Mont Blanc in the Graian Alps, above sea level. The lowest points in the EU are Lammefjorden, Denmark, and Zuidplaspolder, Netherlands, at below sea level. The landscape, climate, and economy of the EU are influenced by its coastline, which is long. In addition to national territories in Europe, there are 32 special territories of members of the European Economic Area, not all of which are part of the EU. The largest by area is Greenland, which is not part of the EU but whose citizens are EU citizens, while the largest by population are the Canary Islands off Africa, which are part of the EU and the Schengen area. French Guiana in South America is part of the EU and the Eurozone, as is Mayotte, north of Madagascar. Climate The climate of the European Union is of a temperate, continental nature, with a maritime climate prevailing on the western coasts and a mediterranean climate in the south. The climate is strongly conditioned by the Gulf Stream, which warms the western region to levels unattainable at similar latitudes on other continents. Western Europe is oceanic, while eastern Europe is continental and dry. Four seasons occur in western Europe, while southern Europe experiences a wet season and a dry season. Southern Europe is hot and dry during the summer months. The heaviest precipitation occurs downwind of water bodies due to the prevailing westerlies, with higher amounts also seen in the Alps. Environment In 1957, when the European Economic Community was founded, it had no environmental policy. Over the past 50 years, an increasingly dense network of legislation has been created, extending to all areas of environmental protection, including air pollution, water quality, waste management, nature conservation, and the control of chemicals, industrial hazards, and biotechnology. According to the Institute for European Environmental Policy, environmental law comprises over 500 Directives, Regulations and Decisions, making environmental policy a core area of European politics. European policy-makers originally increased the EU's capacity to act on environmental issues by defining it as a trade problem. Trade barriers and competitive distortions in the Common Market could emerge due to the different environmental standards in each member state. In subsequent years, the environment became a formal policy area, with its own policy actors, principles and procedures. The legal basis for EU environmental policy was established with the introduction of the Single European Act in 1987. Initially, EU environmental policy focused on Europe. More recently, the EU has demonstrated leadership in global environmental governance, e.g. the role of the EU in securing the ratification and coming into force of the Kyoto Protocol despite opposition from the United States. This international dimension is reflected in the EU's Sixth Environmental Action Programme, which recognises that its objectives can only be achieved if key international agreements are actively supported and properly implemented both at EU level and worldwide. The Lisbon Treaty further strengthened the leadership ambitions. EU law has played a significant role in improving habitat and species protection in Europe, as well as contributing to improvements in air and water quality and waste management. Mitigating climate change is one of the top priorities of EU environmental policy. In 2007, member states agreed that, in the future, 20 per cent of the energy used across the EU must be renewable, and carbon dioxide emissions have to be lower in 2020 by at least 20 per cent compared to 1990 levels. In 2017, the EU emitted 9.1 per cent of global greenhouse-gas emissions. The European Union claims that already in 2018, its GHG emissions were 23% lower than in 1990. The EU has adopted an emissions trading system to incorporate carbon emissions into the economy. The European Green Capital is an annual award given to cities that focuses on the environment, energy efficiency, and quality of life in urban areas to create smart city. In the 2019 elections to the European Parliament, the green parties increased their power, possibly because of the rise of post materialist values. Proposals to reach a zero carbon economy in the European Union by 2050 were suggested in 2018 – 2019. Almost all member states supported that goal at an EU summit in June 2019. The Czech Republic, Estonia, Hungary, and Poland disagreed. In June 2021, the European Union passed a European Climate Law with targets of 55% GHG emissions reduction by 2030 and carbon neutrality by 2050. Also in the same year, the European Union and the United States pledged to cut methane emissions by 30% by 2030. The pledge is considered as a big achievement for climate change mitigation. Economy The gross domestic product (GDP), a measure of economic activity, of the EU was US$16.64 trillion in 2022, around 16.6 percent of the world GDP. This was the third-largest in the world after the United States, and China. There is a significant variation in GDP per capita between and within individual EU states. The difference between the richest and poorest regions (281 NUTS-2 regions of the Nomenclature of Territorial Units for Statistics) ranged, in 2017, from 31 per cent (Severozapaden, Bulgaria) of the EU28 average () to 253 per cent (Luxembourg), or from to . EU member states own the estimated third largest after the United States (trillion) and China (trillion) net wealth in the world, equal to around one sixth (trillion) of the trillion global wealth. Of the top 500 largest corporations in the world measured by revenue in 2010, 161 had their headquarters in the EU. In 2016, unemployment in the EU stood at 8.9 per cent while inflation was at 2.2 per cent, and the account balance at −0.9 per cent of GDP. The average annual net earnings in the European Union was around in 2021. Economic and monetary union The Euro is the official currency in 20 member states of the EU. The creation of a European single currency became an official objective of the European Economic Community in 1969. In 1992, having negotiated the structure and procedures of a currency union, the member states signed the Maastricht Treaty and were legally bound to fulfil the agreed-on rules including the convergence criteria if they wanted to join the monetary union. The states wanting to participate had first to join the European Exchange Rate Mechanism. To prevent the joining states from getting into financial trouble or crisis after entering the monetary union, they were obliged in the Maastricht treaty to fulfil important financial obligations and procedures, especially to show budgetary discipline and a high degree of sustainable economic convergence, as well as to avoid excessive government deficits and limit the government debt to a sustainable level, as agreed in the European Fiscal Pact. Capital Markets Union and financial institutions Free movement of capital is intended to permit movement of investments such as property purchases and buying of shares between countries. Until the drive towards economic and monetary union the development of the capital provisions had been slow. Post-Maastricht there has been a rapidly developing corpus of ECJ judgements regarding this initially neglected freedom. The free movement of capital is unique insofar as it is granted equally to non-member states. The European System of Financial Supervision is an institutional architecture of the EU's framework of financial supervision composed by three authorities: the European Banking Authority, the European Insurance and Occupational Pensions Authority and the European Securities and Markets Authority. To complement this framework, there is also a European Systemic Risk Board under the responsibility of the central bank. The aim of this financial control system is to ensure the economic stability of the EU. Eurozone and banking union In 1999, the currency union started to materialise through introducing a common accounting (virtual) currency in eleven of the member states. In 2002, it was turned into a fully-fledged conventible currency, when euro notes and coins were issued, while the phaseout of national currencies in the eurozone (consisting by then of 12 member states) was initiated. The eurozone (constituted by the EU member states which have adopted the euro) has since grown to 20 countries. The 20 EU member states known collectively as the eurozone have fully implemented the currency union by superseding their national currencies with the euro. The currency union represents 345million EU citizens. The euro is the second largest reserve currency as well as the second most traded currency in the world after the United States dollar. The euro, and the monetary policies of those who have adopted it in agreement with the EU, are under the control of the ECB. The ECB is the central bank for the eurozone, and thus controls monetary policy in that area with an agenda to maintain price stability. It is at the centre of the Eurosystem, which comprehends all the Eurozone national central banks. The ECB is also the central institution of the Banking Union established within the eurozone and manages its Single Supervisory Mechanism. These is also a Single Resolution Mechanism in case of a bank default. Trade As a political entity, the European Union is represented in the World Trade Organization (WTO). Two of the original core objectives of the European Economic Community were the development of a common market, subsequently becoming a single market, and a customs union between its member states. Single market The single market involves the free circulation of goods, capital, people, and services within the EU, The free movement of services and of establishment allows self-employed persons to move between member states to provide services on a temporary or permanent basis. While services account for 60 per cent to 70 per cent of GDP, legislation in the area is not as developed as in other areas. This lacuna has been addressed by the Services in the Internal Market Directive 2006 which aims to liberalise the cross border provision of services. According to the treaty the provision of services is a residual freedom that only applies if no other freedom is being exercised. Customs union The customs union involves the application of a common external tariff on all goods entering the market. Once goods have been admitted into the market they cannot be subjected to customs duties, discriminatory taxes or import quotas, as they travel internally. The non-EU member states of Iceland, Norway, Liechtenstein and Switzerland participate in the single market but not in the customs union. Half the trade in the EU is covered by legislation harmonised by the EU. The European Union Association Agreement does something similar for a much larger range of countries, partly as a so-called soft approach ('a carrot instead of a stick') to influence the politics in those countries. The European Union represents all its members at the World Trade Organization (WTO), and acts on behalf of member states in any disputes. When the EU negotiates trade related agreement outside the WTO framework, the subsequent agreement must be approved by each individual EU member state government. External trade The European Union has concluded free trade agreements (FTAs) and other agreements with a trade component with many countries worldwide and is negotiating with many others. The European Union's services trade surplus rose from $16 billion in 2000 to more than $250 billion in 2018. In 2020, in part due to the COVID-19 pandemic, China became the EU's largest trading partner, displacing the United States. The European Union is the largest exporter in the world and in 2008 was the largest importer of goods and services. Internal trade between the member states is aided by the removal of barriers to trade such as tariffs and border controls. In the eurozone, trade is helped by not having any currency differences to deal with amongst most members. Competition and consumer protection The EU operates a competition policy intended to ensure undistorted competition within the single market. In 2001 the commission for the first time prevented a merger between two companies based in the United States (General Electric and Honeywell) which had already been approved by their national authority. Another high-profile case, against Microsoft, resulted in the commission fining Microsoft over  million following nine years of legal action. Energy The total energy supply of the EU was 59 billion GJ in 2019, about 10.2 per cent of the world total. Approximately three fifths of the energy available in the EU came from imports (mostly of fossil fuels). Renewable energy contributed 18.1 per cent of the EU's total energy supply in 2019, and 11.1 per cent of the final energy consumption. The EU has had legislative power in the area of energy policy for most of its existence; this has its roots in the original European Coal and Steel Community. The introduction of a mandatory and comprehensive European energy policy was approved at the meeting of the European Council in October 2005, and the first draft policy was published in January 2007. The EU has five key points in its energy policy: increase competition in the internal market, encourage investment and boost interconnections between electricity grids; diversify energy resources with better systems to respond to a crisis; establish a new treaty framework for energy co-operation with Russia while improving relations with energy-rich states in Central Asia and North Africa; use existing energy supplies more efficiently while increasing renewable energy commercialisation; and finally increase funding for new energy technologies. In 2007, EU countries as a whole imported 82 per cent of their oil, 57 per cent of their natural gas and 97.48 per cent of their uranium demands. The three largest suppliers of natural gas to the European Union are Russia, Norway and Algeria, that amounted for about three quarters of the imports in 2019. There is a strong dependence on Russian energy that the EU has been attempting to reduce. However, in May 2022, it was reported that the European Union is preparing another sanction against Russia over its invasion of Ukraine. It is expected to target Russian oil, Russian and Belarusian banks, as well as individuals and companies. According to an article by Reuters, two diplomats stated that the European Union may impose a ban on imports of Russian oil by the end of 2022. In May 2022, the EU Commission published the 'RePowerEU' initiative, a €300 billion plan outlining the path towards the end of EU dependence on Russian fossil fuels by 2030 and the acceleration on the clean energy transition. Transport The European Union manages cross-border road, railway, airport and water infrastructure through the Trans-European Transport Network (TEN-T), created in 1990, and the Trans-European Combined Transport network. TEN-T comprises two network layers: the Core Network, which is to be completed by 2030; and the Comprehensive Network, which is to be completed by 2050. The network is currently made up of 9 core corridors: the Baltic–Adriatic Corridor, the North Sea–Baltic Corridor, the Mediterranean Corridor, the Orient/East–Med Corridor, the Scandinavian–Mediterranean Corridor, the Rhine–Alpine Corridor, the Atlantic Corridor, the North Sea–Mediterranean Corridor, and the Rhine–Danube Corridor. Road transportation was organized under the TEN-T by the Trans-European road network. Bundesautobahn 7 is the longest national motorway in the EU at 963 km (598 mi). Maritime transportation is organized under the TEN-T by the Trans-European Inland Waterway network, and the Trans-European Seaport network. European seaports are categorized as international, community, or regional. The Port of Rotterdam is the busiest in the EU, and the world's largest seaport outside of East Asia, located in and near the city of Rotterdam, in the province of South Holland in the Netherlands. The European Maritime Safety Agency (EMSA), founded in 2002 in Lisbon, Portugal, is charged with reducing the risk of maritime accidents, marine pollution from ships and the loss of human lives at sea by helping to enforce the pertinent EU legislation. Air transportation is organized under the TEN-T by the Trans-European Airport network. European airports are categorized as international, community, or regional. The Charles de Gaulle Airport is the busiest in the EU, located in and near the city of Paris, in France. The European Common Aviation Area (ECAA) is a single market in aviation. ECAA agreements were signed on 5 May 2006 in Salzburg, Austria between the EU and some third countries. The ECAA liberalises the air transport industry by allowing any company from any ECAA member state to fly between any ECAA member states airports, thereby allowing a "foreign" airline to provide domestic flights. The Single European Sky (SES) is an initiative that seeks to reform the European air traffic management system through a series of actions carried out in four different levels (institutional, operational, technological and control and supervision) with the aim of satisfying the needs of the European airspace in terms of capacity, safety, efficiency and environmental impact. Civil aviation safety is under the responsibility of the European Union Aviation Safety Agency (EASA). It carries out certification, regulation and standardisation and also performs investigation and monitoring. The idea of a European-level aviation safety authority goes back to 1996, but the agency was only legally established in 2002, and began operating in 2003. Rail transportation is organized under the TEN-T by the Trans-European Rail network, made up of the high-speed rail network and the conventional rail network. The Gare du Nord railway station is the busiest in the EU, located in and near the city of Paris, in France. Rail transport in Europe is being synchronised with the European Rail Traffic Management System (ERTMS) with the goal of greatly enhancing safety, increase efficiency of train transports and enhance cross-border interoperability. This is done by replacing former national signalling equipment and operational procedures with a single new Europe-wide standard for train control and command systems. This system is conducted by the European Union Agency for Railways (ERA). Telecommunications and space Mobile communication roaming charges are abolished throughout the EU, Iceland, Liechtenstein and Norway. The European Union Agency for the Space Programme (EUSPA), headquartered in Prague, Czech Republic, was established in 2021 to manage the European Union Space Programme in order to implement the pre-existing European Space Policy, established on 22 May 2007 between the EU and the European Space Agency (ESA), known collectively as the European Space Council. This was the first common political framework for space activities established by the EU. Each member state has pursued to some extent their own national space policy, though often co-ordinating through the ESA. Günter Verheugen, the European Commissioner for Enterprise and Industry, has stated that even though the EU is "a world leader in the technology, it is being put on the defensive by the United States and Russia and that it only has about a 10-year technological advantage on China and India, which are racing to catch up." Galileo is a global navigation satellite system (GNSS) that went live in 2016, created by the EU through the ESA, operated by the EUSPA, with two ground operations centres in Fucino, Italy, and Oberpfaffenhofen, Germany. The €10 billion project is named after the Italian astronomer Galileo Galilei. One of the aims of Galileo is to provide an independent high-precision positioning system so European political and military authorities do not have to rely on the US GPS, or the Russian GLONASS systems, which could be disabled or degraded by their operators at any time. The European Geostationary Navigation Overlay Service (EGNOS) is a satellite-based augmentation system (SBAS) developed by the ESA and EUROCONTROL. Currently, it supplements the GPS by reporting on the reliability and accuracy of their positioning data and sending out corrections. The system will supplement Galileo in a future version. The Copernicus Programme is the EU's Earth observation programme coordinated and managed by EUSPA in partnership with ESA. It aims at achieving a global, continuous, autonomous, high quality, wide range Earth observation capacity, providing accurate, timely and easily accessible information to, among other things, improve the management of the environment, understand and mitigate the effects of climate change, and ensure civil security. Agriculture and fisheries The Common Agricultural Policy (CAP) is the agricultural policy of the European Union. It implements a system of agricultural subsidies and other programmes. It was introduced in 1962 and has since then undergone several changes to reduce the EEC budget cost (from 73% in 1985 to 37% in 2017) and consider rural development in its aims. It has, however, been criticised on the grounds of its cost and its environmental and humanitarian effects. Likewise, the Common Fisheries Policy (CFP) is the fisheries policy of the European Union. It sets quotas for which member states are allowed to catch each type of fish, as well as encouraging the fishing industry by various market interventions and fishing subsidies. It was introduced in 2009 with the Treaty of Lisbon, which formally enshrined fisheries conservation policy as one of the handful of "exclusive competences" reserved for the European Union. Regional development The five European Structural and Investment Funds are supporting the development of the EU regions, primarily the underdeveloped ones, located mostly in the states of central and southern Europe. Another fund (the Instrument for Pre-Accession Assistance) provides support for candidate members to transform their country to conform to the EU's standard. Demographic transition to a society of ageing population, low fertility-rates and depopulation of non-metropolitan regions is tackled within this policies. Labour The free movement of persons means that EU citizens can move freely between member states to live, work, study or retire in another country. This required the lowering of administrative formalities and recognition of professional qualifications of other states. The EU seasonally adjusted unemployment rate was 6.7 per cent in September 2018. The euro area unemployment rate was 8.1 per cent. Among the member states, the lowest unemployment rates were recorded in the Czech Republic (2.3 per cent), Germany and Poland (both 3.4 per cent), and the highest in Spain (14.9 per cent) and Greece (19.0 in July 2018). The European Union has long sought to mitigate the effects of free markets by protecting workers' rights and preventing social and environmental dumping. To this end it has adopted laws establishing minimum employment and environmental standards. These included the Working Time Directive and the Environmental Impact Assessment Directive. The European Directive about Minimum Wage, which looks to lift minimum wages and strengthen collective bargaining was approved by the European Parliament in September 2022 Social rights and equality The EU has also sought to coordinate the social security and health systems of member states to facilitate individuals exercising free movement rights and to ensure they maintain their ability to access social security and health services in other member states. Since 2019 there has been a European commissioner for equality and the European Institute for Gender Equality has existed since 2007. A Directive on countering gender-based violence has been proposed. In September 2022, a European Care strategy was approved in order to provide "quality, affordable and accessible care services". The European Social Charter is the main body that recognises the social rights of European citizens. In 2020, the first ever European Union Strategy on LGBTIQ equality was approved under Helena Dalli mandate. In December 2021, the commission announced the intention of codifying a union-wide law against LGBT hate crimes. Freedom, security and justice Since the creation of the European Union in 1993, it has developed its competencies in the area of justice and home affairs; initially at an intergovernmental level and later by supranationalism. Accordingly, the union has legislated in areas such as extradition, family law, asylum law, and criminal justice. The EU has also established agencies to co-ordinate police, prosecution and civil litigations across the member states: Europol for police co-operation, CEPOL for training of police forces and the Eurojust for co-operation between prosecutors and courts. It also operates the EUCARIS database of vehicles and drivers, the Eurodac, the European Criminal Records Information System, the European Cybercrime Centre, FADO, PRADO and others. Prohibitions against discrimination have a long standing in the treaties. In more recent years, these have been supplemented by powers to legislate against discrimination based on race, religion, disability, age, and sexual orientation. The treaties declare that the European Union itself is "founded on the values of respect for human dignity, freedom, democracy, equality, the rule of law and respect for human rights, including the rights of persons belonging to minorities ... in a society in which pluralism, non-discrimination, tolerance, justice, solidarity and equality between women and men prevail." By virtue of these powers, the EU has enacted legislation on sexism in the work-place, age discrimination, and racial discrimination. In 2009, the Lisbon Treaty gave legal effect to the Charter of Fundamental Rights of the European Union. The charter is a codified catalogue of fundamental rights against which the EU's legal acts can be judged. It consolidates many rights which were previously recognised by the Court of Justice and derived from the "constitutional traditions common to the member states." The Court of Justice has long recognised fundamental rights and has, on occasion, invalidated EU legislation based on its failure to adhere to those fundamental rights. Signing the European Convention on Human Rights (ECHR) is a condition for EU membership. Previously, the EU itself could not accede to the convention as it is neither a state nor had the competence to accede. The Lisbon Treaty and Protocol 14 to the ECHR have changed this: the former binds the EU to accede to the convention while the latter formally permits it. The EU is independent from the Council of Europe, although they share purpose and ideas, especially on the rule of law, human rights and democracy. Furthermore, the European Convention on Human Rights and European Social Charter, as well as the source of law for the Charter of Fundamental Rights are created by the Council of Europe. The EU has also promoted human rights issues in the wider world. The EU opposes the death penalty and has proposed its worldwide abolition. Abolition of the death penalty is a condition for EU membership. On 19 October 2020, the European Union revealed new plans to create a legal structure to act against human rights violations worldwide. The new plan was expected to provide the European Union with greater flexibility to target and sanction those responsible for serious human rights violations and abuses around the world. Demographics The population of the EU in 2021 was about 447 million people, corresponding to 5.8 per cent of the world population. The population density across the EU was 106 inhabitants per square kilometre, which is more than the world average. It is highest in areas in central and western Europe, sometimes referred to as the "blue banana", while Sweden and Finland in the north are much more sparsely populated. The total population of the EU has been slightly decreasing for several years, contracting by 0.04 per cent in 2021. This is due to a low birth rate of about 1.5 children per woman, less than the world average of 2.3. In total, 4.1 million babies were born in the EU in 2021. Immigration to Europe partially compensates for the natural population decrease. 5.3 per cent of the people residing in the EU are not EU citizens (a person who has the citizenship of a EU member state is automatically also an EU citizen). There were 31 non-EU citizenships that each accounted for at least 1 per cent of non-EU citizens living in the EU, of which the largest were Moroccan, Turkish, Syrian and Chinese. Around 1.9 million people immigrated to one of the EU member states from a non-EU country during 2020, and a total of 956 000 people emigrated from a member state to go to a non-EU country during the same year. Urbanisation More than two thirds (68.2%) of EU inhabitants lived in urban areas in 2020, which is slightly less than the world average. Cities are largely spread out across the EU with a large grouping in and around the Benelux. The EU contains about 40 urban areas with populations of over 1million. With a population of over 13 million, Paris is the largest metropolitan area and the only megacity in the EU. Paris is followed by Madrid, Barcelona, Berlin, the Ruhr, Milan, and Rome, all with a metropolitan population of over 4million. The EU also has numerous polycentric urbanised regions like Rhine-Ruhr (Cologne, Dortmund, Düsseldorf et al.), Randstad (Amsterdam, Rotterdam, The Hague, Utrecht et al.), Frankfurt Rhine-Main (Frankfurt, Wiesbaden, Mainz et al.), the Flemish Diamond (Antwerp, Brussels, Leuven, Ghent et al.) and Upper Silesian area (Katowice, Ostrava et al.). Languages The EU has 24 official languages: Bulgarian, Croatian, Czech, Danish, Dutch, English, Estonian, Finnish, French, German, Greek, Hungarian, Italian, Irish, Latvian, Lithuanian, Maltese, Polish, Portuguese, Romanian, Slovak, Slovene, Spanish, and Swedish. Important documents, such as legislation, are translated into every official language and the European Parliament provides translation for documents and plenary sessions. Most EU institutions use only a handful of working languages: the European Commission conducts its internal business in three procedural languages: English, French, and German; the Court of Justice uses French as the working language, and the European Central Bank conducts its business primarily in English. Even though language policy is the responsibility of member states, EU institutions promote multilingualism among its citizens. The most widely spoken language in the EU is English; the language is spoken by 44 per cent of the population (2016 data) and studied by 95 per cent of school students, although following the withdrawal of the United Kingdom less than 1 per cent of the population speak it natively. German and French are spoken by 36 per cent and 30 per cent of the population. More than half (56 per cent) of EU citizens are able to engage in a conversation in a language other than their mother tongue. Luxembourgish (in Luxembourg) and Turkish (in Cyprus) are the only two national languages that are not official languages of the EU. Catalan, Galician and Basque are not recognised official languages of the EU but have official status in Spain. Therefore, official translations of the treaties are made into them and citizens have the right to correspond with the institutions in these languages. There are about 150 regional and minority languages in the EU, spoken by up to 50 million people. The European Charter for Regional or Minority Languages ratified by most EU states provides general guidelines that states can follow to protect their linguistic heritage. The European Day of Languages is held annually on 26 September and is aimed at encouraging language learning across Europe. Religion The EU has no formal connection to any religion. Article 17 of the Treaty on the Functioning of the European Union recognises the "status under national law of churches and religious associations" as well as that of "philosophical and non-confessional organisations". The preamble to the Treaty on European Union mentions the "cultural, religious and humanist inheritance of Europe". Discussion over the draft texts of the European Constitution and later the Treaty of Lisbon included proposals to mention Christianity or a god, or both, in the preamble of the text, but the idea faced opposition and was dropped. Christians in the EU comprise members of Catholicism (both Roman and Eastern Rite), numerous Protestant denominations (Lutherans, Anglicans, and Reformed forming the bulk of this category), and the Eastern Orthodox Church. In 2009, the EU had an estimated Muslim population of 13 million, and an estimated Jewish population of over a million. The other world religions of Buddhism, Hinduism, and Sikhism are also represented in the EU population. Eurostat's Eurobarometer opinion polls showed in 2005 that 52 per cent of EU citizens believed in a god, 27 per cent in "some sort of spirit or life force", and 18 per cent had no form of belief. Many countries have experienced falling church attendance and membership in recent years. The countries where the fewest people reported a religious belief were Estonia (16 per cent) and the Czech Republic (19 per cent). The most religious countries were Malta (95 per cent, predominantly Catholic) as well as Cyprus and Romania (both predominantly Orthodox) each with about 90 per cent of citizens professing a belief in God. Across the EU, belief was higher among women, older people, those with religious upbringing, those who left school at 15 or 16, and those "positioning themselves on the right of the political scale". Education and research Basic education is an area where the EU's role is limited to supporting national governments. In higher education, the policy was developed in the 1980s in programmes supporting exchanges and mobility. The most visible of these has been the Erasmus Programme, a university exchange programme which began in 1987. In its first 20 years, it supported international exchange opportunities for well over 1.5 million university and college students and became a symbol of European student life. There are similar programmes for school pupils and teachers, for trainees in vocational education and training, and for adult learners in the Lifelong Learning Programme 2007–2013. These programmes are designed to encourage a wider knowledge of other countries and to spread good practices in the education and training fields across the EU. Through its support of the Bologna Process, the EU is supporting comparable standards and compatible degrees across Europe. Scientific development is facilitated through the EU's Framework Programmes, the first of which started in 1984. The aims of EU policy in this area are to co-ordinate and stimulate research. The independent European Research Council allocates EU funds to European or national research projects. EU research and technological framework programmes deal in a number of areas, for example energy where the aim is to develop a diverse mix of renewable energy to help the environment and to reduce dependence on imported fuels. Health Article 35 of the Charter of Fundamental Rights of the European Union affirms that "A high level of human health protection shall be ensured in the definition and implementation of all Union policies and activities". The European Commission's Directorate-General for Health and Consumers seeks to align national laws on the protection of people's health, on the consumers' rights, on the safety of food and other products. All EU and many other European countries offer their citizens a free European Health Insurance Card which, on a reciprocal basis, provides insurance for emergency medical treatment insurance when visiting other participating European countries. A directive on cross-border healthcare aims at promoting co-operation on health care between member states and facilitating access to safe and high-quality cross-border healthcare for European patients. The life expectancy in the EU was 80.1 year at birth in 2021, among the highest in the world and around nine years higher than the world average. In general, life expectancy is lower in Eastern Europe than in Western Europe. In 2018, the EU region with the highest life expectancy was Madrid, Spain at 85.2 years, followed by the Spanish regions of La Rioja and Castilla y León both at 84.3 years, Trentino in Italy at 84.3 years and Île-de-France in France at 84.2 years. Culture Cultural co-operation between member states has been an interest of the European Union since its inclusion as a community competency in the Maastricht Treaty. Actions taken in the cultural area by the EU include the Culture 2000 seven-year programme, the European Cultural Month event, and orchestras such as the European Union Youth Orchestra. The European Capital of Culture programme selects one or more cities in every year to assist the cultural development of that city. Sport Sport is mainly the responsibility of the member states or other international organisations, rather than of the EU. There are some EU policies that have affected sport, such as the free movement of workers, which was at the core of the Bosman ruling that prohibited national football leagues from imposing quotas on foreign players with EU member state citizenship. The Treaty of Lisbon requires any application of economic rules to take into account the specific nature of sport and its structures based on voluntary activity. This followed lobbying by governing organisations such as the International Olympic Committee and FIFA, due to objections over the application of free market principles to sport, which led to an increasing gap between rich and poor clubs. The EU does fund a programme for Israeli, Jordanian, Irish, and British football coaches, as part of the Football 4 Peace project. Symbols The flag of Europe consists of a circle of 12 golden stars on a blue background. Originally designed in 1955 for the Council of Europe, the flag was adopted by the European Communities, the predecessors of the present European Union, in 1986. The Council of Europe gave the flag a symbolic description in the following terms, though the official symbolic description adopted by the EU omits the reference to the "Western world": United in Diversity was adopted as the motto of the union in 2000, having been selected from proposals submitted by school pupils. Since 1985, the flag day of the union has been Europe Day, on 9 May (the date of the 1950 Schuman declaration). The anthem of the EU is an instrumental version of the prelude to the Ode to Joy, the 4th movement of Ludwig van Beethoven's ninth symphony. The anthem was adopted by European Community leaders in 1985 and has since been played on official occasions. Besides naming the continent, the Greek mythological figure of Europa has frequently been employed as a personification of Europe. Known from the myth in which Zeus seduces her in the guise of a white bull, Europa has also been referred to in relation to the present union. Statues of Europa and the bull decorate several of the EU's institutions and a portrait of her is seen on the 2013 series of euro banknotes. The bull is, for its part, depicted on all residence permit cards. Charles the Great, also known as Charlemagne () and later recognised as Pater Europae ("Father of Europe"), has a symbolic relevance to Europe. The commission has named one of its central buildings in Brussels after Charlemagne and the city of Aachen has since 1949 awarded the Charlemagne Prize to champions of European unification. Since 2008, the organisers of this prize, in conjunction with the European Parliament, have awarded the Charlemagne Youth Prize in recognition of similar efforts led by young people. Media Media freedom is a fundamental right that applies to all member states of the European Union and its citizens, as defined in the EU Charter of Fundamental Rights as well as the European Convention on Human Rights. Within the EU enlargement process, guaranteeing media freedom is named a "key indicator of a country's readiness to become part of the EU". The majority of media in the European Union are national-orientated, although some EU-wide media focusing on European affairs have emerged since the early 1990s, such as Euronews, Eurosport, EUobserver, EURACTIV or Politico Europe. Arte is a public Franco-German TV network that promotes programming in the areas of culture and the arts. 80 per cent of its programming are provided in equal proportion by the two member companies, while the remainder is being provided by the European Economic Interest Grouping ARTE GEIE and the channel's European partners. The MEDIA Programme of the European Union has supported the European popular film and audiovisual industries since 1991. It provides support for the development, promotion and distribution of European works within Europe and beyond. Influence The European Union has had a significant positive economic effect on most member states. According to a 2019 study of the member states who joined from 1973 to 2004, "without European integration, per capita incomes would have been, on average, approximately 10 per cent lower in the first ten years after joining the EU." Greece was the exception reported by the study, which analysed up to 2008, "to avoid confounding effects from the global financial crisis". A 2021 study in the Journal of Political Economy found that the 2004 enlargement had aggregate beneficial economic effects on all groups in both the old and new member states. The largest winners were the new member states, in particular unskilled labour in the new member states. The European Union has contributed to peace in Europe, in particular by pacifying border disputes, and to the spread of democracy, especially by encouraging democratic reforms in aspiring Eastern European member states after the collapse of the USSR. Scholar Thomas Risse wrote in 2009, "there is a consensus in the literature on Eastern Europe that the EU membership perspective had a huge anchoring effects for the new democracies." However, R. Daniel Kelemen argues that the EU has proved beneficial to leaders who are overseeing democratic backsliding, as the EU is reluctant to intervene in domestic politics, gives authoritarian governments funds which they can use to strengthen their regimes, and because freedom of movement within the EU allows dissenting citizens to leave their backsliding countries. At the same time, the union provides an external constraint that prevents soft authoritarian regimes from progressing into hard dictatorships. See also Outline of the European Union Special territories of members of the European Economic Area List of country groupings List of multilateral free-trade agreements Euroscepticism Pan-European nationalism Brexit withdrawal agreement EU–UK Trade and Cooperation Agreement African Union Notes References Citations Bibliography Further reading Excerpt and text search. External links EUROPA – official web portal Historical Archives of the European Union European Union. The World Factbook. Central Intelligence Agency. The European Union: Questions and Answers. Congressional Research Service. European Economic Area 1993 establishments in Europe Articles containing video clips Confederations G20 nations International organizations based in Europe International political organizations Organisations based in Brussels Organizations awarded Nobel Peace Prizes Organizations established in 1993 Political organizations based in Europe Political systems Supranational unions Trade blocs United Nations General Assembly observers
9433
https://en.wikipedia.org/wiki/Spain%20in%20Flames
Spain in Flames
Spain in Flames is a 1937 compilation film made by Helen van Dongen during the Spanish Civil War. Hal Erickson has written that the film "... is remarkable in its willingness to offer both sides of the conflict -- though its sympathies are firmly with the Loyalists." The film consists of two parts. The first, "The Fight for Freedom", was based on film footage from a Spanish government documentary Spain and the Fight for Freedom. A foreword by the then Spanish Ambassador to the United States, Fernando de los Ríos, began one of the film's screenings in New York in 1937. The second part, "They Shalt Not Pass", was based on a short film No Pasaran! done by the Artkino Film Company of the Soviet Union, where van Dongen was working at the time the film was made. John Dos Passos narrated parts of the film, and the commentary was written by Dos Passos, Ernest Hemingway, Archibald MacLeish, and Prudencio de Pareda. Erickson writes that, "The horrendous images of battlefield carnage, not to mention the close-ups of suffering and dying Spanish children, still pack a wallop when seen today." Later, Hemingway, Dos Passos, Lillian Hellman and others founded the company Contemporary Historians, which produced another film called The Spanish Earth (1937), directed by Joris Ivens and edited by van Dongen. Spain in Flames was banned in New Brunswick, New Jersey and Waterbury, Connecticut. A screening of the film, accompanied by a speech from Granville Hicks, was also banned in Provincetown, Massachusetts. See also The Spanish Earth (1937) España 1936 (1937) References Further reading 1937 films Anti-fascist propaganda films Spanish Civil War films Compilation films American anthology films Features based on short films American black-and-white films American propaganda films
9620
https://en.wikipedia.org/wiki/Education%20reform
Education reform
Education reform is the name given to the goal of changing public education. The meaning and education methods have changed through debates over what content or experiences result in an educated individual or an educated society. Historically, the motivations for reform have not reflected the current needs of society. A consistent theme of reform includes the idea that large systematic changes to educational standards will produce social returns in citizens' health, wealth, and well-being. As part of the broader social and political processes, the term education reform refers to the chronology of significant, systematic revisions made to amend the educational legislation, standards, methodology, and policy affecting a nation's public school system to reflect the needs and values of contemporary society.18th century, classical education instruction from an in-home personal tutor, hired at the family's expense, was primarily a privilege for children from wealthy families. Innovations such as encyclopedias, public libraries, and grammar schools all aimed to relieve some of the financial burden associated with the expenses of the classical education model. Motivations during the Victorian era emphasized the importance of self-improvement. Victorian education focused on teaching commercially valuable topics, such as modern languages and mathematics, rather than classical liberal arts subjects, such as Latin, art, and history. Motivations for education reformists like Horace Mann and his proponents focused on making schooling more accessible and developing a robust state-supported common school system. John Dewey, an early 20th-century reformer, focused on improving society by advocating for a scientific, pragmatic, or democratic principle-based curriculum. Whereas Maria Montessori incorporated humanistic motivations to "meet the needs of the child". In historic Prussia, a motivation to foster national unity led to formal education concentrated on teaching national language literacy to young children, resulting in Kindergarten. The history of educational pedagogy in the United States has ranged from teaching literacy and proficiency of religious doctrine to establishing cultural literacy, assimilating immigrants into a democratic society, producing a skilled labor force for the industrialized workplace, preparing students for careers, and competing in a global marketplace. Education inequality is also a motivation for education reform, seeking to address problems of a community. Motivations for education reform Education reform, in general, implies a continual effort to modify and improve the institution of education. Over time, as the needs and values of society change, attitudes towards public education change. As a social institution, education plays an integral role in the process of socialization. "Socialization is broadly composed of distinct inter- and intra-generational processes. Both involve the harmonization of an individual's attitudes and behaviors with that of their socio-cultural milieu." Educational matrices mean to reinforce those socially acceptable informal and formal norms, values, and beliefs that individuals need to learn in order to be accepted as good, functioning, and productive members of their society. Education reform is the process of constantly renegotiating and restructuring the educational standards to reflect the ever-evolving contemporary ideals of social, economic, and political culture. Reforms can be based on bringing education into alignment with a society's core values. Reforms that attempt to change a society's core values can connect alternative education initiatives with a network of other alternative institutions. Education reform has been pursued for a variety of specific reasons, but generally most reforms aim at redressing some societal ills, such as poverty-, gender-, or class-based inequities, or perceived ineffectiveness. Current education trends in the United States represent multiple achievement gaps across ethnicities, income levels, and geographies. As McKinsey and Company reported in a 2009 analysis, "These educational gaps impose on the United States the economic equivalent of a permanent national recession." Reforms are usually proposed by thinkers who aim to redress societal ills or institute societal changes, most often through a change in the education of the members of a class of people—the preparation of a ruling class to rule or a working class to work, the social hygiene of a lower or immigrant class, the preparation of citizens in a democracy or republic, etc. The idea that all children should be provided with a high level of education is a relatively recent idea, and has arisen largely in the context of Western democracy in the 20th century. The "beliefs" of school districts are optimistic that quite literally "all students will succeed", which in the context of high school graduation examination in the United States, all students in all groups, regardless of heritage or income will pass tests that in the introduction typically fall beyond the ability of all but the top 20 to 30 percent of students. The claims clearly renounce historical research that shows that all ethnic and income groups score differently on all standardized tests and standards based assessments and that students will achieve on a bell curve. Instead, education officials across the world believe that by setting clear, achievable, higher standards, aligning the curriculum, and assessing outcomes, learning can be increased for all students, and more students can succeed than the 50 percent who are defined to be above or below grade level by norm referenced standards. States have tried to use state schools to increase state power, especially to make better soldiers and workers. This strategy was first adopted to unify related linguistic groups in Europe, including France, Germany and Italy. Exact mechanisms are unclear, but it often fails in areas where populations are culturally segregated, as when the U.S. Indian school service failed to suppress Lakota and Navaho, or when a culture has widely respected autonomous cultural institutions, as when the Spanish failed to suppress Catalan. Many students of democracy have desired to improve education in order to improve the quality of governance in democratic societies; the necessity of good public education follows logically if one believes that the quality of democratic governance depends on the ability of citizens to make informed, intelligent choices, and that education can improve these abilities. Politically motivated educational reforms of the democratic type are recorded as far back as Plato in The Republic. In the United States, this lineage of democratic education reform was continued by Thomas Jefferson, who advocated ambitious reforms partly along Platonic lines for public schooling in Virginia. Another motivation for reform is the desire to address socio-economic problems, which many people see as having significant roots in lack of education. Starting in the 20th century, people have attempted to argue that small improvements in education can have large returns in such areas as health, wealth and well-being. For example, in Kerala, India in the 1950s, increases in women's health were correlated with increases in female literacy rates. In Iran, increased primary education was correlated with increased farming efficiencies and income. In both cases some researchers have concluded these correlations as representing an underlying causal relationship: education causes socio-economic benefits. In the case of Iran, researchers concluded that the improvements were due to farmers gaining reliable access to national crop prices and scientific farming information. History Classical education As taught from the 18th to the 19th century, Western classical education curriculums focused on concrete details like "Who?", "What?", "When?", "Where?". Unless carefully taught, large group instruction naturally neglects asking the theoretical "Why?" and "Which?" questions that can be discussed in smaller groups. Classical education in this period also did not teach local (vernacular) languages and culture. Instead, it taught high-status ancient languages (Greek and Latin) and their cultures. This produced odd social effects in which an intellectual class might be more loyal to ancient cultures and institutions than to their native vernacular languages and their actual governing authorities. 18th century Child-study Jean-Jacques Rousseau, father of the Child Study Movement, centered the child as an object of study. In Emile: Or, On Education, Rousseau's principal work on education lays out an educational program for a hypothetical newborn's education through adulthood. Rousseau provided a dual critique of the educational vision outlined in Plato's Republic and that of his society in contemporary Europe. He regarded the educational methods contributing to the child's development; he held that a person could either be a man or a citizen. While Plato's plan could have brought the latter at the expense of the former, contemporary education failed at both tasks. He advocated a radical withdrawal of the child from society and an educational process that utilized the child's natural potential and curiosity, teaching the child by confronting them with simulated real-life obstacles and conditioning the child through experience rather intellectual instruction. Rousseau ideas were rarely implemented directly, but influenced later thinkers, particularly Johann Heinrich Pestalozzi and Friedrich Wilhelm August Fröbel, the inventor of the kindergarten. National identity European and Asian nations regard education as essential to maintaining national, cultural, and linguistic unity. In the late 18th century (~1779), Prussia instituted primary school reforms expressly to teach a unified version of the national language, "Hochdeutsch". One significant reform was kindergarten whose purpose was to have the children participate in supervised activities taught by instructors who spoke the national language. The concept embraced the idea that children absorb new language skills more easily and quickly when they are young The current model of kindergarten is reflective of the Prussian model. In other countries, such as the Soviet Union, France, Spain, and Germany, the Prussian model has dramatically improved reading and math test scores for linguistic minorities. 19th century England In the 19th century, before the advent of government-funded public schools, Protestant organizations established Charity Schools to educate the lower social classes. The Roman Catholic Church and governments later adopted the model. Designed to be inexpensive, Charity schools operated on minimal budgets and strived to serve as many needy children as possible. This led to the development of grammar schools, which primarily focused on teaching literacy, grammar, and bookkeeping skills so that the students could use books as an inexpensive resource to continue their education. Grammar was the first third of the then-prevalent system of classical education.. Educators Joseph Lancaster and Andrew Bell developed the monitorial system, also known as "mutual instruction" or the "Bell–Lancaster method". Their contemporary, educationalist and writer Elizabeth Hamilton, suggested that in some important aspects the method had been "anticipated" by the Belfast schoolmaster David Manson. In the 1760s Manson had developed a peer-teaching and monitoring system within the context of what he called a "play school" that dispensed with "the discipline of the rod". (More radically, Manson proposed the "liberty of each [child] to take the quantity [of lessons] agreeable to his inclination"). Lancaster, an impoverished Quaker during the early 19th century in London and Bell at the Madras School of India developed this model independent of one another. However, by design, their model utilizes more advanced students as a resource to teach the less advanced students; achieving student-teacher ratios as small as 1:2 and educating more than 1000 students per adult. The lack of adult supervision at the Lancaster school resulted in the older children acting as disciplinary monitors and taskmasters. To provide order and promote discipline the school implemented a unique internal economic system, inventing a currency called a Scrip. Although the currency was worthless in the outside world, it was created at a fixed exchange rate from a student's tuition and student's could use scrip to buy food, school supplies, books, and other items from the school store. Students could earn scrip through tutoring. To promote discipline, the school adopted a work-study model. Every job of the school was bid-for by students, with the largest bid winning. However, any student tutor could auction positions in his or her classes to earn scrip. The bids for student jobs paid for the adult supervision. Lancaster promoted his system in a piece called Improvements in Education that spread widely throughout the English-speaking world. Lancaster schools provided a grammar-school education with fully developed internal economies for a cost per student near $40 per year in 1999 U.S. dollars. To reduce cost and motivated to save up scrip, Lancaster students rented individual pages of textbooks from the school library instead of purchasing the textbook. Student's would read aloud their pages to groups. Students commonly exchanged tutoring and paid for items and services with receipts from down tutoring. The schools did not teach submission to orthodox Christian beliefs or government authorities. As a result, most English-speaking countries developed mandatory publicly paid education explicitly to keep public education in "responsible" hands. These elites said that Lancaster schools might become dishonest, provide poor education, and were not accountable to established authorities. Lancaster's supporters responded that any child could cheat given the opportunity, and that the government was not paying for the education and thus deserved no say in their composition. Though motivated by charity, Lancaster claimed in his pamphlets to be surprised to find that he lived well on the income of his school, even while the low costs made it available to the most impoverished street children. Ironically, Lancaster lived on the charity of friends in his later life. Modern reformist Although educational reform occurred on a local level at various points throughout history, the modern notion of education reform is tied with the spread of compulsory education. Economic growth and the spread of democracy raised the value of education and increased the importance of ensuring that all children and adults have access to free, high-quality, effective education. Modern education reforms are increasingly driven by a growing understanding of what works in education and how to go about successfully improving teaching and learning in schools. However, in some cases, the reformers' goals of "high-quality education" has meant "high-intensity education", with a narrow emphasis on teaching individual, test-friendly subskills quickly, regardless of long-term outcomes, developmental appropriateness, or broader educational goals. Horace Mann In the United States, Horace Mann (1796 – 1859) of Massachusetts used his political base and role as Secretary of the Massachusetts State Board of Education to promote public education in his home state and nationwide. Advocating a substantial public investment be made in education, Mann and his proponents developed a strong system of state supported common schools.. His crusading style attracted wide middle class support. Historian Ellwood P. Cubberley asserts: No one did more than he to establish in the minds of the American people the conception that education should be universal, non-sectarian, free, and that its aims should be social efficiency, civic virtue, and character, rather than mere learning or the advancement of sectarian ends. In 1852, Massachusetts passed a law making education mandatory. This model of free, accessible education spread throughout the country and in 1917 Mississippi was the final state to adopt the law. John Dewey John Dewey, a philosopher and educator based in Chicago and New York, helped conceptualize the role of American and international education during the first four decades of the 20th century. An important member of the American Pragmatist movement, he carried the subordination of knowledge to action into the educational world by arguing for experiential education that would enable children to learn theory and practice simultaneously; a well-known example is the practice of teaching elementary physics and biology to students while preparing a meal. He was a harsh critic of "dead" knowledge disconnected from practical human life. Dewey criticized the rigidity and volume of humanistic education, and the emotional idealizations of education based on the child-study movement that had been inspired by Rousseau and those who followed him. Dewey understood that children are naturally active and curious and learn by doing. Dewey's understanding of logic is presented in his work "Logic, the Theory of Inquiry" (1938). His educational philosophies were presented in "My Pedagogic Creed", The School and Society, The Child and Curriculum, and Democracy and Education (1916). Bertrand Russell criticized Dewey's conception of logic, saying "What he calls "logic" does not seem to me to be part of logic at all; I should call it part of psychology." Dewey left the University of Chicago in 1904 over issues relating to the Dewey School. Dewey's influence began to decline in the time after the Second World War and particularly in the Cold War era, as more conservative educational policies came to the fore. Administrative progressives The form of educational progressivism which was most successful in having its policies implemented has been dubbed "administrative progressivism" by historians. This began to be implemented in the early 20th century. While influenced particularly in its rhetoric by Dewey and even more by his popularizers, administrative progressivism was in its practice much more influenced by the Industrial Revolution and the concept economies of scale. The administrative progressives are responsible for many features of modern American education, especially American high schools: counseling programs, the move from many small local high schools to large centralized high schools, curricular differentiation in the form of electives and tracking, curricular, professional, and other forms of standardization, and an increase in state and federal regulation and bureaucracy, with a corresponding reduction of local control at the school board level. (Cf. "State, federal, and local control of education in the United States", below) (Tyack and Cuban, pp. 17–26) These reforms have since become heavily entrenched, and many today who identify themselves as progressives are opposed to many of them, while conservative education reform during the Cold War embraced them as a framework for strengthening traditional curriculum and standards. More recent methods, instituted by groups such as the think tank Reform's education division, and S.E.R. have attempted to pressure the government of the U.K. into more modernist educational reform, though this has met with limited success. Public school reform in the United States In the United States, public education is characterized as "any federally funded primary or secondary school, administered to some extent by the government, and charged with educating all citizens. Although there is typically a cost to attend some public higher education institutions, they are still considered part of public education." Colonial America In what would become the United States, the first public school was established in Boston, Massachusetts, on April 23, 1635. Puritan schoolmaster Philemon Pormont led instruction at the Boston Latin School. During this time, post-secondary education was a commonly utilized tool to distinguish one's social class and social status. Access to education was the "privilege of white, upper-class, Christian male children" in preparation for university education in ministry. In colonial America, to maintain Puritan religious traditions, formal and informal education instruction focused on teaching literacy. All colonists needed to understand the written language on some fundamental level in order to read the Bible and the colony's written secular laws. Religious leaders recognized that each person should be "educated enough to meet the individual needs of their station in life and social harmony." The first compulsory education laws were passed in Massachusetts between 1642 and 1648 when religious leaders noticed not all parents were providing their children with proper education. These laws stated that all towns with 50 or more families were obligated to hire a schoolmaster to teach children reading, writing, and basic arithmetic."In 1642 the General Court passed a law that required heads of households to teach all their dependents — apprentices and servants as well as their own children — to read English or face a fine. Parents could provide the instruction themselves or hire someone else to do it. Selectmen were to keep 'a vigilant eye over their brethren and neighbors,' young people whose education was neglected could be removed from their parents or masters."The 1647 law eventually led to establishing publicly funded district schools in all Massachusetts towns, although, despite the threat of fines, compliance and quality of public schools were less than satisfactory."Many towns were 'shamefully neglectful' of children's education. In 1718 '...by sad experience, it is found that many towns that not only are obliged by law, but are very able to support a grammar school, yet choose rather to incur and pay the fine or penalty than maintain a grammar school."When John Adams drafted the Massachusetts Constitution in 1780, he included provisions for a comprehensive education law that guaranteed public education to "all" citizens. However, access to formal education in secondary schools and colleges was reserved for free, white males. During the 17th and 18th centuries, females received little or no formal education except for home learning or attending Dame Schools. Likewise, many educational institutions maintained a policy of refusing to admit Black applicants. The Virginia Code of 1819 outlawed teaching enslaved people to read or write. Post-revolution Soon after the American Revolution, early leaders, like Thomas Jefferson and John Adams, proposed the creation of a more "formal and unified system of publicly funded schools" to satiate the need to "build and maintain commerce, agriculture and shipping interests". Their concept of free public education was not well received and did not begin to take hold on until the 1830s. However, in 1790, evolving socio-cultural ideals in the Commonwealth of Pennsylvania led to the first significant and systematic reform in education legislation that mandated economic conditions would not inhibit a child's access to education:"Constitution of the Commonwealth of Pennsylvania – 1790 ARTICLE VII Section I. The legislature shall, as soon as conveniently may be, provide, by law, for the establishment of schools throughout the state, in such manner that the poor may be taught gratis." Reconstruction and the American Industrial Revolution During Reconstruction, from 1865 to 1877, African Americans worked to encourage public education in the South. With the U.S. Supreme Court decision in Plessy v. Ferguson, which held that "segregated public facilities were constitutional so long as the black and white facilities were equal to each other", this meant that African American children were legally allowed to attend public schools, although these schools were still segregated based on race. However, by the mid-twentieth century, civil rights groups would challenge racial segregation. During the second half of the nineteenth century (1870 and 1914), America's Industrial Revolution refocused the nation's attention on the need for a universally accessible public school system. Inventions, innovations, and improved production methods were critical to the continued growth of American manufacturing. To compete in the global economy, an overwhelming demand for literate workers that possessed practical training emerged. Citizens argued, "educating children of the poor and middle classes would prepare them to obtain good jobs, thereby strengthen the nation's economic position." Institutions became an essential tool in yielding ideal factory workers with sought-after attitudes and desired traits such as dependability, obedience, and punctuality. Vocationally oriented schools offered practical subjects like shop classes for students who were not planning to attend college for financial or other reasons. Not until the latter part of the 19th century did public elementary schools become available throughout the country. Although, it would be longer for children of color, girls, and children with special needs to attain access free public education. Mid 20th and early 21st century (United states) Civil rights reform Systemic bias remained a formidable barrier. From the 1950s to the 1970s, many of the proposed and implemented reforms in U.S. education stemmed from the civil rights movement and related trends; examples include ending racial segregation, and busing for the purpose of desegregation, affirmative action, and banning of school prayer. In the early 1950s, most U.S. public schools operated under a legally sanctioned racial segregation system. Civil Rights reform movements sought to address the biases that ensure unequal distribution of academic resources such as school funding, qualified and experienced teachers, and learning materials to those socially excluded communities. In the early 1950s, the NAACP lawyers brought class-action lawsuits on behalf of black schoolchildren and their families in Kansas, South Carolina, Virginia, and Delaware, petitioning court orders to compel school districts to let black students attend white public schools. Finally, in 1954, the U.S. Supreme Court rejected that framework with Brown v. Board of Education and declared state-sponsored segregation of public schools unconstitutional. In 1964, Title VI of the Civil Rights Act (Public Law 88-352) "prohibited discrimination on the basis of race, color, and national origin in programs and activities receiving federal financial assistance." Educational institutions could now utilize public funds to implement in-service training programs to assist teachers and administrators in establishing desegregation plans. In 1965, the Higher Education Act (HEA) (Public Law 89–329) authorizes federal aid for postsecondary students. The Elementary and Secondary Education Act of 1965 (ESEA) (Public Law 89-313) represents the federal government's commitment to providing equal access to quality education; including those children from low-income families, limited English proficiency, and other minority groups. This legislation had positive retroactive implications for Historically Black Colleges and Universities, more commonly known as HBCUs."The Higher Education Act of 1965, as amended, defines an HBCU as: "…any historically black college or university that was established prior to 1964, whose principal mission was, and is, the education of black Americans, and that is accredited by a nationally recognized accrediting agency or association determined by the Secretary [of Education] to be a reliable authority as to the quality of training offered or is, according to such an agency or association, making reasonable progress toward accreditation."Known as the Bilingual Education Act, Title VII of ESEA (Public Law 90-247), offered federal aid to school districts to provide bilingual instruction for students with limited English speaking ability. The Education Amendments of 1972 (Public Law 92-318, 86 Stat. 327) establishes the Education Division in the U.S. Department of Health, Education, and Welfare and the National Institute of Education. Title IX of the Education Amendments of 1972 states, "No person in the United States shall, on the basis of sex, be excluded from participation in, be denied the benefits of, or be subjected to discrimination under any education program or activity receiving Federal financial assistance." Equal Educational Opportunities Act of 1974 (Public Law 93-380) - Civil Rights Amendments to the Elementary and Secondary Education Act of 1965:"Title I: Bilingual Education Act - Authorizes appropriations for carrying out the provisions of this Act. Establishes, in the Office of Education, an Office of Bilingual Education through which the Commissioner of Education shall carry out his functions relating to bilingual education. Authorizes appropriations for school nutrition and health services, correction education services, and ethnic heritage studies centers. Title II: Equal Educational Opportunities and the Transportation of Students: Equal Educational Opportunities Act - Provides that no state shall deny equal educational opportunity to an individual on account of his or her race, color, sex, or national origin by means of specified practices... Title IV: Consolidation of Certain Education Programs: Authorizes appropriations for use in various education programs including libraries and learning resources, education for use of the metric system of measurement, gifted and talented children programs, community schools, career education, consumers' education, women's equity in education programs, and arts in education programs. Community Schools Act - Authorizes the Commissioner to make grants to local educational agencies to assist in planning, establishing, expanding, and operating community education programs Women's Educational Equity Act - Establishes the Advisory Council on Women's Educational Programs and sets forth the composition of such Council. Authorizes the Commissioner of Education to make grants to, and enter into contracts with, public agencies, private nonprofit organizations, and individuals for activities designed to provide educational equity for women in the United States. Title V: Education Administration: Family Educational Rights and Privacy Act (FERPA)- Provides that no funds shall be made available under the General Education Provisions Act to any State or local educational agency or educational institution which denies or prevents the parents of students to inspect and review all records and files regarding their children. Title VII: National Reading Improvement Program: Authorizes the Commissioner to contract with State or local educational agencies for the carrying out by such agencies, in schools having large numbers of children with reading deficiencies, of demonstration projects involving the use of innovative methods, systems, materials, or programs which show promise of overcoming such reading deficiencies."In 1975, The Education for All Handicapped Children Act (Public Law 94-142) ensured that all handicapped children (age 3-21) receive a "free, appropriate public education" designed to meet their special needs. 1980-1989: A Nation at Risk During the 1980s, some of the momentum of education reform moved from the left to the right, with the release of A Nation at Risk, Ronald Reagan's efforts to reduce or eliminate the United States Department of Education. "[T]he federal government and virtually all state governments, teacher training institutions, teachers' unions, major foundations, and the mass media have all pushed strenuously for higher standards, greater accountability, more "time on task," and more impressive academic results". Per the shift in educational motivation, families sought institutional alternatives, including "charter schools, progressive schools, Montessori schools, Waldorf schools, Afrocentric schools, religious schools - or home school instruction in their communities." In 1984 President Reagan enacted the Education for Economic Security Act (Public Law 98-377) In 1989, the Child Development and Education Act of 1989 (Public Law 101-239) authorized funds for Head Start Programs to include child care services. In the latter half of the decade, E. D. Hirsch put forth an influential attack on one or more versions of progressive education. Advocating an emphasis on "cultural literacy"—the facts, phrases, and texts. See also Uncommon Schools. 1990-2000: standards-based education model In 1994, the land grant system was expanded via the Elementary and Secondary Education Act to include tribal colleges. Most states and districts in the 1990s adopted outcome-based education (OBE) in some form or another. A state would create a committee to adopt standards, and choose a quantitative instrument to assess whether the students knew the required content or could perform the required tasks. In 1992 The National Commission on Time and Learning, Extension (Public Law 102–359) revise funding for civic education programs and those educationally disadvantaged children." In 1994 the Improving America's Schools Act (IASA) (Public Law 103-382) reauthorized the Elementary and Secondary Education Act of 1965; amended as The Eisenhower Professional Development Program; IASA designated Title I funds for low income and otherwise marginalized groups; i.e., females, minorities, individuals with disabilities, individuals with limited English proficiency (LEP). By tethering federal funding distributions to student achievement, IASA meant use high stakes testing and curriculum standards to hold schools accountable for their results at the same level as other students. The Act significantly increased impact aid for the establishment of the Charter School Program, drug awareness campaigns, bilingual education, and technology. In 1998 The Charter School Expansion Act (Public Law 105-278) amended the Charter School Program, enacted in 1994. 2001-2015: No Child Left Behind Consolidated Appropriations Act of 2001 (Public Law 106-554) appropriated funding to repair educational institution's buildings as well as repair and renovate charter school facilities, reauthorized the Even Start program, and enacted the Children's Internet Protection Act. The standards-based National Education Goals 2000, set by the U.S. Congress in the 1990s, were based on the principles of outcomes-based education. In 2002, the standards-based reform movement culminated as the No Child left Behind Act of 2001 (Public Law 107-110) where achievement standard were set by each individual state. This federal policy was active until 2015 in the United States . An article released by CBNC.com said a principal Senate Committee will take into account legislation that reauthorizes and modernizes the Carl D. Perkins Act. President George Bush approved this statute in 2006 on August 12, 2006. This new bill will emphasize the importance of federal funding for various Career and Technical (CTE) programs that will better provide learners with in-demand skills. Pell Grants are specific amount of money is given by the government every school year for disadvantaged students who need to pay tuition fees in college. At present, there are many initiatives aimed at dealing with these concerns like innovative cooperation between federal and state governments, educators, and the business sector. One of these efforts is the Pathways to Technology Early College High School (P-TECH). This six-year program was launched in cooperation with IBM, educators from three cities in New York, Chicago, and Connecticut, and over 400 businesses. The program offers students in high school and associate programs focusing on the STEM curriculum. The High School Involvement Partnership, private and public venture, was established through the help of Northrop Grumman, a global security firm. It has given assistance to some 7,000 high school students (juniors and seniors) since 1971 by means of one-on-one coaching as well as exposure to STEM areas and careers. 2016-2021: Every Student Succeeds Act The American Reinvestment and Recovery Act, enacted in 2009, reserved more than $85 billion in public funds to be used for education. The 2009 Council of Chief State School Officers and the National Governors Association launch the Common Core State Standards Initiative. In 2012 the Obama administration launched the Race to the Top competition aimed at spurring K–12 education reform through higher standards."The Race to the Top – District competition will encourage transformative change within schools, targeted toward leveraging, enhancing, and improving classroom practices and resources. The four key areas of reform include: Development of rigorous standards and better assessments Adoption of better data systems to provide schools, teachers, and parents with information about student progress Support for teachers and school leaders to become more effective Increased emphasis and resources for the rigorous interventions needed to turn around the lowest-performing schools"In 2015, under the Obama administration, many of the more restrictive elements that were enacted under No Child Left Behind (NCLB, 2001), were removed in the Every Student Succeeds Act (ESSA, 2015) which limits the role of the federal government in school liability. Every Student Succeeds Act (Public Law 114-95) reformed educational standards by "moving away from such high stakes and assessment based accountability models" and focused on assessing student achievement from a holistic approach by utilizing qualitative measures. Some argue that giving states more authority can help prevent considerable discrepancies in educational performance across different states. ESSA was approved by former President Obama in 2015 which amended and empowered the Elementary and Secondary Education Act of 1965. The Department of Education has the choice to carry out measures in drawing attention to said differences by pinpointing lowest-performing state governments and supplying information on the condition and progress of each state on different educational parameters. It can also provide reasonable funding along with technical aid to help states with similar demographics collaborate in improving their public education programs. Social and emotional learning: strengths-based education model This uses a methodology that values purposeful engagement in activities that turn students into self-reliant and efficient learners. Holding on to the view that everyone possesses natural gifts that are unique to one's personality (e.g. computational aptitude, musical talent, visual arts abilities), it likewise upholds the idea that children, despite their inexperience and tender age, are capable of coping with anguish, able to survive hardships, and can rise above difficult times. Trump administration In 2017, Betsy DeVos was instated as the 11th Secretary of Education. A strong proponent of school choice, school voucher programs, and charter schools, DeVos was a much-contested choice as her own education and career had little to do with formal experience in the US education system. In a Republican-dominated senate, she received a 50–50 vote - a tie that was broken by Vice President Mike Pence. Prior to her appointment, DeVos received a BA degree in business economics from Calvin College in Grand Rapids, Michigan and she served as chairman of an investment management firm, The Windquest Group. She supported the idea of leaving education to state governments under the new K-12 legislation. DeVos cited the interventionist approach of the federal government to education policy following the signing of the ESSA. The primary approach to that rule has not changed significantly. Her opinion was that the education movement populist politics or populism encouraged reformers to commit promises which were not very realistic and therefore difficult to deliver. On July 31, 2018, President Donald Trump signed the Strengthening Career and Technical Education for the 21st Century Act (HR 2353) The Act reauthorized the Carl D. Perkins Career and Technical Education Act, a $1.2 billion program modified by the United States Congress in 2006. A move to change the Higher Education Act was also deferred. The legislation enacted on July 1, 2019, replaced the Carl D. Perkins Career and Technical Education (Perkins IV) Act of 2006. Stipulations in Perkins V enables school districts to make use of federal subsidies for all students' career search and development activities in the middle grades as well as comprehensive guidance and academic mentoring in the upper grades. At the same time, this law revised the meaning of "special populations" to include homeless persons, foster youth, those who left the foster care system, and children with parents on active duty in the United States armed forces. Barriers to reform Education inequalities facing students of color Another factor to consider in education reform is that of equity and access. Contemporary issues in the United States regarding education faces a history of inequalities that come with consequences for education attainment across different social groups. Racial and socioeconomic class segregation A history of racial, and subsequently class, segregation in the U.S. resulted from practices of law. Residential segregation is a direct result of twentieth century policies that separated by race using zoning and redlining practices, in addition to other housing policies, whose effects continue to endure in the United States. These neighborhoods that have been segregated de jure—by force of purposeful public policy at the federal, state, and local levels—disadvantage people of color as students must attend school near their homes. With the inception of the New Deal between 1933 and 1939, and during and following World War II, federally funded public housing was explicitly racially segregated by the local government in conjunction with federal policies through projects that were designated for Whites or Black Americans in the South, Northeast, Midwest, and West. Following an ease on the housing shortage post-World War II, the federal government subsidized the relocation of Whites to suburbs. The Federal Housing and Veterans Administration constructed such developments on the East Coast in towns like Levittown on Long Island, New Jersey, Pennsylvania, and Delaware. On the West Coast, there was Panorama City, Lakewood, Westlake, and Seattle suburbs developed by Bertha and William Boeing. As White families left for the suburbs, Black families remained in public housing and were explicitly placed in Black neighborhoods. Policies such as public housing director, Harold Ickes', "neighborhood composition rule" maintained this segregation by establishing that public housing must not interfere with pre-existing racial compositions of neighborhoods. Federal loan guarantees were given to builders who adhered to the condition that no sales were made to Black families and each deed prohibited re-sales to Black families, what the Federal Housing Administration (FHA) described as an "incompatible racial element". In addition, banks and savings intuitions refused loans to Black families in White suburbs and Black families in Black neighborhoods. In the mid-twentieth century, urban renewal programs forced low-income black residents to reside in places farther from universities, hospitals, or business districts and relocation options consisted of public housing high-rises and ghettos. This history of de jure segregation has impacted resource allocation for public education in the United States, with schools continuing to be segregated by race and class. Low-income White students are more likely than Black students to be integrated into middle-class neighborhoods and less likely to attend schools with other predominantly disadvantaged students. Students of color disproportionately attend underfunded schools and Title I schools in environments entrenched in environmental pollution and stagnant economic mobility with limited access to college readiness resources. According to research, schools attended by primarily Hispanic or African American students often have high turnover of teaching staff and are labeled high-poverty schools, in addition to having limited educational specialists, less available extracurricular opportunities, greater numbers of provisionally licensed teachers, little access to technology, and buildings that are not well maintained. With this segregation, more local property tax is allocated to wealthier communities and public schools' dependence on local property taxes has led to large disparities in funding between neighboring districts. The top 10% of wealthiest school districts spend approximately ten times more per student than the poorest 10% of school districts. Racial wealth gap This history of racial and socioeconomic class segregation in the U.S. has manifested into a racial wealth divide. With this history of geographic and economic segregation, trends illustrate a racial wealth gap that has impacted educational outcomes and its concomitant economic gains for minorities. Wealth or net worth—the difference between gross assets and debt—is a stock of financial resources and a significant indicator of financial security that offers a more complete measure of household capability and functioning than income. Within the same income bracket, the chance of completing college differs for White and Black students. Nationally, White students are at least 11% more likely to complete college across all four income groups. Intergenerational wealth is another result of this history, with White college-educated families three times as likely as Black families to get an inheritance of $10,000 or more. 10.6% of White children from low-income backgrounds and 2.5% of Black children from low-income backgrounds reach the top 20% of income distribution as adults. Less than 10% of Black children from low-income backgrounds reach the top 40%. Access to early childhood education These disadvantages facing students of color are apparent early on in early childhood education. By the age of five, children of color are impacted by opportunity gaps indicated by poverty, school readiness gap, segregated low-income neighborhoods, implicit bias, and inequalities within the justice system as Hispanic and African American boys account for as much as 60% of total prisoners within the incarceration population. These populations are also more likely to experience adverse childhood experiences (ACEs). High-quality early care and education are less accessible to children of color, particularly African American preschoolers as findings from the National Center for Education Statistics show that in 2013, 40% of Hispanic and 36% White children were enrolled in learning center-based classrooms rated as high, while 25% of African American children were enrolled in these programs. 15% of African American children attended low ranking center-based classrooms. In home-based settings, 30% of White children and over 50% of Hispanic and African American children attended low rated programs. Contemporary issues (United States) Overview In the first decade of the 21st century, several issues are salient in debates over further education reform: Longer school day or school year After-school tutoring Charter schools, school choice, or school vouchers Smaller class sizes Improved teacher quality Improved training Higher credential standards Generally higher pay to attract more qualified applicants Performance bonuses ("merit pay") Firing low-performing teachers Internet and computer access in schools Track and reduce drop-out rate Track and reduce absenteeism English-only vs. bilingual education Mainstreaming or fully including students with special educational needs, rather than placing them in separate special schools Content of curriculum standards and textbooks What to teach, at what age, and to which students. Discussion points include the age at which children should learn to read, and the primary mathematical subject that is taught to adolescents – algebra, or statistics or personal finances. Funding, neglected infrastructure, and adequacy of educational supplies Student rights Education inequalities facing students of color Private interest in American charter schools Charter schools public independent institutions in which both the cost and risk are fully funded by the taxpayers. Some charter schools are nonprofit in name only and are structured in ways that individuals and private enterprises connected to them can make money. Other charter schools are for-profit. In many cases, the public is largely unaware of this rapidly changing educational landscape, the debate between public and private/market approaches, and the decisions that are being made that affect their children and communities. Critics have accused for-profit entities, (education management organizations, EMOs) and private foundations such as the Bill and Melinda Gates Foundation, the Eli and Edythe Broad Foundation, and the Walton Family Foundation of funding Charter school initiatives to undermine public education and turn education into a "Business Model" which can make a profit. In some cases a school's charter is held by a non-profit that chooses to contract all of the school's operations to a third party, often a for-profit, CMO. This arrangement is defined as a vendor-operated school, (VOS). School choice Economists such as Nobel laureate Milton Friedman advocate school choice to promote excellence in education through competition and choice. A competitive "market" for schools eliminates the need to otherwise attempt a workable method of accountability for results. Public education vouchers permit guardians to select and pay any school, public or private, with public funds currently allocated to local public schools. The theory is that children's guardians will naturally shop for the best schools, much as is already done at college level. Though appealing in theory, many reforms based on school choice have led to slight to moderate improvements—which some teachers' union members see as insufficient to offset the decreased teacher pay and job security. For instance, New Zealand's landmark reform in 1989, during which schools were granted substantial autonomy, funding was devolved to schools, and parents were given a free choice of which school their children would attend, led to moderate improvements in most schools. It was argued that the associated increases in inequity and greater racial stratification in schools nullified the educational gains. Others, however, argued that the original system created more inequity (due to lower income students being required to attend poorer performing inner city schools and not being allowed school choice or better educations that are available to higher income inhabitants of suburbs). Instead, it was argued that the school choice promoted social mobility and increased test scores especially in the cases of low income students. Similar results have been found in other jurisdictions. Though discouraging, the merely slight improvements of some school choice policies often seems to reflect weaknesses in the way that choice is implemented rather than a failure of the basic principle itself. Teacher tenure Critics of teacher tenure claim that the laws protect ineffective teachers from being fired, which can be detrimental to student success. Tenure laws vary from state to state, but generally they set a probationary period during which the teacher proves themselves worthy of the lifelong position. Probationary periods range from one to three years. Advocates for tenure reform often consider these periods too short to make such an important decision; especially when that decision is exceptionally hard to revoke. Due process restriction protect tenured teachers from being wrongfully fired; however these restrictions can also prevent administrators from removing ineffective or inappropriate teachers. A 2008 survey conducted by the US Department of Education found that, on average, only 2.1% of teachers are dismissed each year for poor performance. In October 2010 Apple Inc. CEO Steve Jobs had a consequential meeting with U.S. President Barack Obama to discuss U.S. competitiveness and the nation's education system. During the meeting Jobs recommended pursuing policies that would make it easier for school principals to hire and fire teachers based on merit. In 2012 tenure for school teachers was challenged in a California lawsuit called Vergara v. California. The primary issue in the case was the impact of tenure on student outcomes and on equity in education. On June 10, 2014, the trial judge ruled that California's teacher tenure statute produced disparities that " shock the conscience" and violate the equal protection clause of the California Constitution. On July 7, 2014, U.S. Secretary of Education Arne Duncan commented on the Vergara decision during a meeting with President Barack Obama and representatives of teacher's unions. Duncan said that tenure for school teachers "should be earned through demonstrated effectiveness" and should not be granted too quickly. Specifically, he criticized the 18-month tenure period at the heart of the Vergara case as being too short to be a "meaningful bar." Funding levels According to a 2005 report from the OECD, the United States is tied for first place with Switzerland when it comes to annual spending per student on its public schools, with each of those two countries spending more than $11,000 (in U.S. currency). Despite this high level of funding, U.S. public schools lag behind the schools of other rich countries in the areas of reading, math, and science. A further analysis of developed countries shows no correlation between per student spending and student performance, suggesting that there are other factors influencing education. Top performers include Singapore, Finland and Korea, all with relatively low spending on education, while high spenders including Norway and Luxembourg have relatively low performance. One possible factor is the distribution of the funding. In the US, schools in wealthy areas tend to be over-funded while schools in poorer areas tend to be underfunded. These differences in spending between schools or districts may accentuate inequalities, if they result in the best teachers moving to teach in the most wealthy areas. The inequality between districts and schools led to 23 states instituting school finance reform based on adequacy standards that aim to increase funding to low-income districts. A 2018 study found that between 1990 and 2012, these finance reforms led to an increase in funding and test scores in the low income districts; which suggests finance reform is effective at bridging inter-district performance inequalities. It has also been shown that the socioeconomic situation of the students family has the most influence in determining success; suggesting that even if increased funds in a low income area increase performance, they may still perform worse than their peers from wealthier districts. Starting in the early 1980s, a series of analyses by Eric Hanushek indicated that the amount spent on schools bore little relationship to student learning. This controversial argument, which focused attention on how money was spent instead of how much was spent, led to lengthy scholarly exchanges. In part the arguments fed into the class size debates and other discussions of "input policies." It also moved reform efforts towards issues of school accountability (including No Child Left Behind) and the use of merit pay and other incentives. There have been studies that show smaller class sizes and newer buildings (both of which require higher funding to implement) lead to academic improvements. It should also be noted that many of the reform ideas that stray from the traditional format require greater funding. According to a 1999 article, William J. Bennett, former U.S. Secretary of Education, argued that increased levels of spending on public education have not made the schools better, citing the following statistics: Internationally Education for All Education 2030 Agenda refers to the global commitment of the Education for All movement to ensure access to basic education for all. It is an essential part of the 2030 Agenda for Sustainable Development. The roadmap to achieve the Agenda is the Education 2030 Incheon Declaration and Framework for Action, which outlines how countries, working with UNESCO and global partners, can translate commitments into action. The United Nations, over 70 ministers, representatives of member-countries, bilateral and multilateral agencies, regional organizations, academic institutions, teachers, civil society, and the youth supported the Framework for Action of the Education 2030 platform. The Framework was described as the outcome of continuing consultation to provide guidance for countries in implementing this Agenda. At the same time, it mobilizes various stakeholders in the new education objectives, coordination, implementation process, funding, and review of Education 2030. Thailand In 1995, the minister of education, Sukavich Rangsitpol, launched a series of education reforms in 1995 with the intention of the education reform is to realize the potential of Thai people to develop themselves for a better quality of life and to develop the nation for a peaceful co-existence in the global community. The follow-on Reform Program of 1996 is built around four major improvements: improving the physical state of schools upgrading the quality of teachers reforming learning and teaching methods streamlining administration School-based management (SBM) in Thailand implemented in 1997 in the course of a reform aimed at overcoming a profound crisis in the educación system. According to UNESCO, Thailand education reform has led to the following results: The educational budget increased from 133 billion baht in 1996 to 163 billion baht in 1997 (22.5% increase) Since 1996, first grade students have been taught English as a second or foreign language and computer literacy. Professional advancement from teacher level 6 to teacher level 7 without having to submit academic work for consideration was approved by the Thai government. Free 12 years education for all children provided by the government. The Eighth Thailand’s National Social and Economic Development Plan was also written to support the implemented of education reform program. This program was later added to the 1997 Constitution of Thailand and gave access to all citizens. World Bank report that after the 1997 Asian financial crisis Income in the northeast, the poorest part of Thailand, has risen by 46 percent from 1998 to 2006. Nationwide poverty fell from 21.3 to 11.3 percent. Learning crisis The learning crisis is the reality that while the majority of children around the world attend school, a large proportion of them are not learning. A World Bank study found that "53 percent of children in low- and middle-income countries cannot read and understand a simple story by the end of primary school." While schooling has increased rapidly over the last few decades, learning has not followed suit. Many practitioners and academics call for education system reform in order to address the learning needs of all children. Digital education The movement to use computers more in education naturally includes many unrelated ideas, methods, and pedagogies since there are many uses for digital computers. For example, the fact that computers are naturally good at math leads to the question of the use of calculators in math education. The Internet's communication capabilities make it potentially useful for collaboration, and foreign language learning. The computer's ability to simulate physical systems makes it potentially useful in teaching science. More often, however, debate of digital education reform centers around more general applications of computers to education, such as electronic test-taking and online classes. Another viable addition to digital education has been blended learning. In 2009, over 3 million K-12 students took an online course, compared to 2000 when 45,000 took an online course. Blended learning examples include pure online, blended, and traditional education. Research results show that the most effective learning takes place in a blended format. This allows children to view the lecture ahead of time and then spend class time practicing, refining, and applying what they have previously learned. The idea of creating artificial intelligence led some computer scientists to believe that teachers could be replaced by computers, through something like an expert system; however, attempts to accomplish this have predictably proved inflexible. The computer is now more understood to be a tool or assistant for the teacher and students. Harnessing the richness of the Internet is another goal. In some cases classrooms have been moved entirely online, while in other instances the goal is more to learn how the Internet can be more than a classroom. Web-based international educational software is under development by students at New York University, based on the belief that current educational institutions are too rigid: effective teaching is not routine, students are not passive, and questions of practice are not predictable or standardized. The software allows for courses tailored to an individual's abilities through frequent and automatic multiple intelligences assessments. Ultimate goals include assisting students to be intrinsically motivated to educate themselves, and aiding the student in self-actualization. Courses typically taught only in college are being reformatted so that they can be taught to any level of student, whereby elementary school students may learn the foundations of any topic they desire. Such a program has the potential to remove the bureaucratic inefficiencies of education in modern countries, and with the decreasing digital divide, help developing nations rapidly achieve a similar quality of education. With an open format similar to Wikipedia, any teacher may upload their courses online and a feedback system will help students choose relevant courses of the highest quality. Teachers can provide links in their digital courses to webcast videos of their lectures. Students will have personal academic profiles and a forum will allow students to pose complex questions, while simpler questions will be automatically answered by the software, which will bring you to a solution by searching through the knowledge database, which includes all available courses and topics. The 21st century ushered in the acceptance and encouragement of internet research conducted on college and university campuses, in homes, and even in gathering areas of shopping centers. Addition of cyber cafes on campuses and coffee shops, loaning of communication devices from libraries, and availability of more portable technology devices, opened up a world of educational resources. Availability of knowledge to the elite had always been obvious, yet provision of networking devices, even wireless gadget sign-outs from libraries, made availability of information an expectation of most persons. Cassandra B. Whyte researched the future of computer use on higher education campuses focusing on student affairs. Though at first seen as a data collection and outcome reporting tool, the use of computer technology in the classrooms, meeting areas, and homes continued to unfold. The sole dependence on paper resources for subject information diminished and e-books and articles, as well as online courses, were anticipated to become increasingly staple and affordable choices provided by higher education institutions according to Whyte in a 2002 presentation. Digitally "flipping" classrooms is a trend in digital education that has gained significant momentum. Will Richardson, author and visionary for the digital education realm, points to the not-so-distant future and the seemingly infinite possibilities for digital communication linked to improved education. Education on the whole, as a stand-alone entity, has been slow to embrace these changes. The use of web tools such as wikis, blogs, and social networking sites is tied to increasing overall effectiveness of digital education in schools. Examples exist of teacher and student success stories where learning has transcended the classroom and has reached far out into society. The media has been instrumental in pushing formal educational institutions to become savvier in their methods. Additionally, advertising has been (and continues to be) a vital force in shaping students and parents thought patterns. Technology is a dynamic entity that is constantly in flux. As time presses on, new technologies will continue to break paradigms that will reshape human thinking regarding technological innovation. This concept stresses a certain disconnect between teachers and learners and the growing chasm that started some time ago. Richardson asserts that traditional classroom's will essentially enter entropy unless teachers increase their comfort and proficiency with technology. Administrators are not exempt from the technological disconnect. They must recognize the existence of a younger generation of teachers who were born during the Digital Age and are very comfortable with technology. However, when old meets new, especially in a mentoring situation, conflict seems inevitable. Ironically, the answer to the outdated mentor may be digital collaboration with worldwide mentor webs; composed of individuals with creative ideas for the classroom. See also Anti-schooling activism Blab school Block scheduling Certificate of Initial Mastery Criterion-referenced test Educational philosophies Excellence and equity Female education High school graduation examination Higher-order thinking Inquiry-based Science Learning crisis Learning environment Learning space Merit pay Multiculturalism Political correctness Project-based learning Special Assistance Program Student-centered learning Sudbury model democratic schools Sudbury Valley School Teaching for social justice University reform Web literacy References Sources Further reading Comer, J.P. (1997). Waiting for a Miracle: Why Schools Can't Solve Our Problems- and How We Can. New York: Penguin Books. Cuban, L. (2003). Why Is It So Hard to Get Good Schools? New York: Teachers College, Columbia University. Darling-Hammond, Linda. (1997) The Right to Learn: A Blueprint for Creating Schools that Work. Jossey-Bass. Dewey, J. and Dewey, E. (1915). Schools of To-morrow. New York: E.P. Dutton and Company. Gatto, John Taylor (1992). Dumbing Us Down: The Hidden Curriculum of Compulsory Schooling. Canada: New Society Publishers. Glazek, S.D. and Sarason, S.B. (2007). Productive Learning: Science, Art, and Einstein's Relativity in Education Reform. New York: Sage Publications, Inc. Goodland, J.I. and Anderson, R.H. (1959 and 1987). The Nongraded Elementary School. New York: Harcourt, Brace and Company. James, Laurie. (1994) Outrageous Questions: Legacy of Bronson Alcott and America's One-Room Schools New York. Katz, M.B. (1971). Class, Bureaucracy, and Schools: The Illusion of Educational Change in America. New York: Praeger Publishers. Kliebard, Herbert. (1987) The Struggle for the American Curriculum. New York : Routledge & Kegan Paul. Kohn, A. (1999). The Schools Our Children Deserve: Moving Beyond Traditional Classrooms and 'Tougher Standards. Boston: Houghton Mifflin Co. Murphy, J.H. and Beck, L.G. (1995). School-Based Management as School Reform: Taking Stock. Thousand Oaks, CA: Corwin Press, Inc. Ogbu, J.U. (1978). Minority Education and Caste: The American System in Cross-Cultural Perspective. New York: Academic Press. Ravitch, D. (1988). The Great School Wars: A History of the New York City Public Schools. New York: Basic Books, Inc. Sarason, S.B. (1996). Revisiting 'The Culture of the School and the Problem of Change. New York: Teachers College Press. Sarason, S.B. (1990). The Predictable Failure of Educational Reform: Can We Change Course Before Its Too Late? San Francisco: Josey-Bass, Inc. Sizer, T.R. (1984). Horace's Compromise: The Dilemma of the American High School. Boston: Houghton Mifflin Company. Tough, Paul. (2008). Whatever It Takes: Geoffrey Canada's Quest to Change Harlem and America. New York: Houghton Mifflin Company. Tough, Paul. (2012). How Children Succeed. New York: Houghton Mifflin Company. Tyack, David and Cuban, Larry. (1995) Tinkering Toward Utopia: A Century of Public School Reform. Cambridge, MA: Harvard University Press. Zwaagstra, Michael; Clifton, Rodney; and Long, John. (2010) What's Wrong with Our Schools: and How We Can Fix Them. Rowman & Littlefield. External links Education issues History of education
9760
https://en.wikipedia.org/wiki/Eschatology
Eschatology
Eschatology (; ) concerns expectations of the end of the present age, human history, or the world itself. The end of the world or end times is predicted by several world religions (both Abrahamic and non-Abrahamic), which teach that negative world events will reach a climax. Belief that the end of the world is imminent is known as apocalypticism, and over time has been held both by members of mainstream religions and by doomsday cults. In the context of mysticism, the term refers metaphorically to the end of ordinary reality and to reunion with the divine. Various religions treat eschatology as a future event prophesied in sacred texts or in folklore. The Abrahamic religions maintain a linear cosmology, with end-time scenarios containing themes of transformation and redemption. In later Judaism, the term "end of days" makes reference to the Messianic Age and includes an in-gathering of the exiled Jewish diaspora, the coming of the Messiah, the resurrection of the righteous, and the world to come. Some forms of Christianity depict the end time as a period of tribulation that precedes the second coming of Christ, who will face the rise of the Antichrist along with his power structure and false prophets, and usher in the Kingdom of God. In Islam, the Day of Judgment is preceded by the appearance of the Masīḥ ad-Dajjāl, and followed by the descending of ʿĪsā (Jesus), which shall triumph over the false Messiah or Antichrist; his defeat will lead to a sequence of events that will end with the sun rising from the west and the beginning of the Qiyāmah (Judgment Day). Dharmic religions tend to have more cyclical worldviews, with end-time eschatologies characterized by decay, redemption, and rebirth (though some believe transitions between cycles are relatively uneventful). In Hinduism, the end time occurs when Kalki, the final incarnation of Vishnu, descends atop a white horse and brings an end to the current Kali Yuga, completing a cycle that starts again with the regeneration of the world. In Buddhism, the Buddha predicted his teachings would be forgotten after 5,000 years, followed by turmoil. It says a bodhisattva named Maitreya will appear and rediscover the teachings of the Buddha Dharma, and that the ultimate destruction of the world will then come through seven suns. Since the development of the concept of deep time in the 18th century the calculation of the estimated age of planet Earth, scientific discourse about end times has considered the ultimate fate of the universe. Theories have included the Big Rip, Big Crunch, Big Bounce, and Big Freeze (heat death). Social and scientific commentators also worry about global catastrophic risks and scenarios that could result in human extinction. Etymology The word "eschatology" arises from the Ancient Greek term (éschatos), meaning "last", and -logy, meaning "the study of", and first appeared in English around 1844. The Oxford English Dictionary defines eschatology as "the part of theology concerned with death, judgment, and the final destiny of the soul and of humankind". Linear cosmology Judaism The main tenets of modern Jewish eschatology, in no particular order, include: God will redeem Israel from the captivity that began during the Babylonian Exile in a new Exodus. God will return the Jewish people to the Land of Israel. God will restore the House of David and the Temple in Jerusalem. God will raise up a regent from the House of David, the Jewish Messiah, to lead the Jewish people and the world and to usher in an age of justice and peace, the Messianic Age. Non-Jews will recognize that the God of Israel is the only true god. God will resurrect the dead. God will create a new heaven and earth. Judaism usually refers to the end times as the "end of days" (aḥarit ha-yamim, אחרית הימים), a phrase that appears several times in the Tanakh. The end times are addressed in the Book of Daniel and in numerous other prophetic passages in the Hebrew scriptures, and also in the Talmud, particularly Tractate Avodah Zarah. The idea of a Messianic Age, an era of global peace and knowledge of the Creator, has a prominent place in Jewish thought, and is incorporated as part of the end of days. A well-known passage from the Book of Isaiah describes this future condition of the world: "They shall beat their swords into plowshares and their spears into pruning hooks; nation will not lift sword against nation and they will no longer study warfare" (2:4). Maimonides (1135–1204) further describes the Messianic Era in the Mishneh Torah: "And at that time there will be no hunger or war, no jealousy or rivalry. For the good will be plentiful, and all delicacies available as dust. The entire occupation of the world will be only to know God;... the people Israel will be of great wisdom; they will perceive the esoteric truths and comprehend their Creator's wisdom as is the capacity of man. As it is written (Isaiah 11:9): 'For the earth shall be filled with the knowledge of God, as the waters cover the sea.'" Kabbalah In Kabbalah, the Zohar maintains that the seven days of the week, based on the seven days of creation, correspond to the seven millennia of creation. The seventh day of the week, the Shabbat day of rest, corresponds to the seventh millennium, the age of universal rest, or the Messianic Era. The seventh millennium begins with the year 6000 AM, and is the latest time the Messiah can come. A number of early and late Jewish scholars have written in support of this, including the Ramban, Isaac Abarbanel, Abraham Ibn Ezra, Rabbeinu Bachya, the Vilna Gaon, the Lubavitcher Rebbe, the Ramchal, Aryeh Kaplan and Rebbetzin Esther Jungreis. Zoroastrianism Frashokereti is the Zoroastrian doctrine of a final renovation of the universe when evil will be destroyed, and everything else will then be in perfect unity with God (Ahura Mazda). The doctrinal premises are: Good will eventually prevail over evil. Creation, initially perfectly good, was subsequently corrupted by evil. The world will ultimately be restored to the perfection it had at the time of creation. The "salvation for the individual depended on the sum of [that person's] thoughts, words and deeds, and there could be no intervention, whether compassionate or capricious, by any divine being to alter this". Thus each human bears the responsibility for the fate of his own soul, and simultaneously shares in the responsibility for the fate of the world. Zoroastrian eschatology is considered one of the oldest in recorded history. The birth of its founder, Zoroaster, is unknown, with scholarly dates ranging from sixth century BCE to 5,500 years earlier. Pliny the Elder even suggests there were two Zoroasters. However, with beliefs paralleling and possibly predating the framework of the major Abrahamic faiths, a fully developed concept of the end of the world was not established in Zoroastrianism until 500 BCE. The Bahman Yasht describes: At the end of thy tenth hundredth winter, the sun is more unseen and more spotted; the year, month, and day are shorter; and the earth is more barren; and the crop will not yield the seed. And men become more deceitful and more given to vile practices. They will have no gratitude. Honorable wealth will proceed to those of perverted faith. And a dark cloud makes the whole sky night, and it will rain more noxious creatures than water. A battle between the righteous and wicked will be followed by the Frashokereti. On earth, the Saoshyant will arrive as the final savior of mankind, and bring about the resurrection of the dead. The yazatas Airyaman and Atar will melt the metal in the hills and mountains, which will flow as lava across the earth and all mankind, both the living and resurrected, will be required to wade through it. Ashavan will pass through the molten river as if it were warm milk, but the sinful will burn. It will then flow down to hell, where it will annihilate Angra Mainyu and the last vestiges of wickedness. The righteous will partake of the parahaoma, which will confer immortality upon them. Humanity will become like the Amesha Spentas, living without food, hunger, thirst, weapons or injury. Bodies will become so light as to cast no shadow. All humanity will speak a single language, and belong to a single nation with no borders. All will share a single purpose and goal, joining with Ahura Mazda for a perpetual and divine exaltation. Gnosticism The Gnostic codex On the Origin of the World states that during what is called the consummation of the age, the Sun and Moon will become dark as the stars change their ordinary course. Kings will make war with each other, and thunder will cause the world to be shaken. The corrupt Archons will mourn. The sea will be troubled by fighting of the kings who became drunk from the flaming sword. Finally, great thunder will come from Sophia, the woman in the firmament above the forces of Chaos. She will cast the corrupt gods into the abyss where they will fight each other until only their chief Yaldabaoth remains and destroys himself. Next the heavens of the Archons will collapse on each other before the Earth sinks into the abyss. Light will cover the darkness and eliminate it then form into something greater than anything that ever existed before. The source of the darkness will dissolve, and the deficiency will be taken from its root. Those who were not perfected in the unconceived one will receive glories in their realms and kingdoms of the immortals, but those who were will enter a kingless realm. All will be judged according to their deeds and gnosis. Christianity Christian eschatology is the study concerned with the ultimate destiny of the individual soul and of the entire created order, based primarily upon biblical texts within the Old and New Testaments. Christian eschatological research looks to study and discuss matters such as the nature of the divine and the divine nature of Jesus Christ, death and the afterlife, Heaven and Hell, the Second Coming of Jesus, the resurrection of the dead, the rapture, the Tribulation, millennialism, the end of the world, the Last Judgment, and the New Heaven and New Earth in the world to come. Eschatological passages occur in many places in the Bible, in both the Old and the New Testaments. In the Old Testament, apocalyptic eschatology can be found notably in Isaiah 24–27, Isaiah 56–66, Joel, Zechariah 9–14 as well as in the closing chapters of Daniel, and in Ezekiel. In the New Testament, applicable passages include Matthew 24, Mark 13, the parable of "The Sheep and the Goats" and the Book of Revelation—Revelation often occupies a central place in Christian eschatology. The Second Coming of Christ is the central event in Christian eschatology within the broader context of the fullness of the Kingdom of God. Most Christians believe that death and suffering will continue to exist until Christ's return. There are, however, various views concerning the order and significance of other eschatological events. The Book of Revelation stands at the core of much of Christian eschatology. The study of Revelation is usually divided into four interpretative methodologies or hermeneutics: The Futurist approach treats the Book of Revelation mostly as unfulfilled prophecy taking place in some yet undetermined future. The Preterist approach interprets Revelation chiefly as having had prophetic fulfillment in the past, principally in the events of the first century CE. The Historicist approach places Revelation within the context of history, identifying figures and passages in Revelation with major historical people and events. This view was commonly held by the early Christian church, then among the predecessors to Protestantism, such as John Wycliffe, Joachim of Fiore and later by the majority of Protestant Reformers, such as Martin Luther, John Calvin, and John Wesley. Further supporters of this view included Isaac Newton (1642-1727), among others. The Idealist approach sees the events of Revelation as neither past nor future actualities, but as purely symbolic accounts, dealing with the ongoing struggle and ultimate triumph of good over evil. Date First-century Christians believed Jesus would return during their lifetime. When the converts of Paul in Thessalonica were persecuted by the Roman Empire, they believed the end of days to be imminent. Scholarly consensus would hold that Jesus, and following him the early Christians, understood the endtime as being imminent. While some who believe in the literal interpretation of the Bible insist the prediction of dates or times is futile, others believe Jesus foretold signs of the end of days. The precise time, however, will come like a "thief in the night" (). They may also refer to in which Jesus is quoted as saying: Great Tribulation In the New Testament, Jesus refers to this period preceding the end times as the "Great Tribulation" (), "Affliction" (), and "days of vengeance" (). The Book of Matthew describes the devastation: The resulting chaos will affect pregnancies, newborns, and a scourge will spread throughout the flesh, save for the elect. The vivid imagery of this section is repeated closely in . The Gospel of Luke describes a complete unraveling of the social fabric, with widespread calamity and war: In the Book of Revelation, the "great tribulation" (Rev. 7:14b) refers to a time of affliction upon God's people. Catholicism and Orthodoxy The Profession of Faith addresses Catholic beliefs concerning the last days. Catholicism adheres to the amillennial school of thought, promoted by Augustine of Hippo in his work The City of God. Protestantism Contemporary use of the term End Times has evolved from literal belief in Christian millennialism. In this tradition, Biblical apocalypse is believed to be imminent, with various current events as omens of impending Armageddon. These beliefs have been put forward by the Adventist movement (Millerites), Jehovah's Witnesses, and dispensational premillennialists. In 1918 a group of eight, well-known preachers produced the London Manifesto, warning of an imminent second coming of Christ shortly after the 1917 liberation of Jerusalem by the British. Millennialists and Amillennialists Protestants are divided between Millennialists and Amillennialists. Millennialists concentrate on the issue of whether the true believers will see the Great Tribulation or be removed from it by what is referred to as a Pre-Tribulation rapture. Amillennialists believe the end times encompass the time from Christ's ascension to the last day, and maintain that the mention of the "thousand years" in the Book of Revelation is meant to be taken metaphorically (i.e., not literally), a view which continues to cause divisions within Protestant Christianity. There is a range of eschatological belief in Protestant Christianity. Christian premillennialists who believe the end times are occurring now, are usually specific about timelines that climax in the end of the world. For some, Israel, the European Union, or the United Nations are seen as major players whose roles were foretold in scripture. Within dispensational premillennialist writing, there is the belief that Christians will be summoned to Heaven by Christ at the rapture, occurring before a Great Tribulation prophesied in Matthew 24–25; Mark 13 and Luke 21. The Tribulation is described in the Book of Revelation. "End times" may also refer to the passing of an age or long period in the relationship between man and God. Adherents to this view cite the Second Epistle to Timothy and draw analogies to the late twentieth and early twenty-first centuries. Post-Exilic Hebrew books of prophecy such as Daniel and Ezekiel are given new interpretations in this Christian tradition, while apocalyptic forecasts appear in the Judeo-Christian Sibylline Oracles which include the Book of Revelation ascribed to John, the apocryphal Apocalypse of Peter, and the Second Book of Esdras. Fundamentalists Most fundamentalist Christians anticipate biblical prophecy to be literally fulfilled. They see current wars, natural disaster and famine as the birth pangs which Jesus described in Matthew and Mark . They believe mankind began in the garden of Eden, and point to the Valley of Megiddo as the place where the current world system will terminate, after which the Messiah will rule for 1,000 years. Adventists and Millerites Religious movements which expect that the second coming of Christ will be a cataclysmic event are generally called adventism. These have arisen throughout the Christian era, but were particularly common after the Protestant Reformation. Emanuel Swedenborg considered the second coming to be symbolic, and to have occurred in 1757. Along with others, he developed a religious system around the second coming of Christ, disclosed by new prophecy or special revelation not described in the Bible. The Millerites are diverse religious groups which similarly rely upon a special gift of interpretation for predicting the second coming. The difference between the 19th-century Millerite and adventist movements and contemporary prophecy is that William Miller and his followers, based on biblical interpretation, predicted the time of the Second Coming to have occurred in 1844. Contemporary writing of end time has suggested the timetable will be triggered by future wars and moral catastrophe, and that this time of tribulation is close at hand. Seventh-day Adventists believe biblical prophecy to foretell an end time scenario in which the United States works in conjunction with the Catholic Church to mandate worship on a day other than the true Sabbath, Saturday, as prescribed in the Ten Commandments (Exodus 20:8–11). This will bring about a situation where one must choose for or against the Bible as the will of God. Preterists Another view of the end times is preterism. It distinguishes the time of the end from the end of time. Preterists believe the term last days (or Time of the End) refers to, neither the last days of the Earth, nor the last days of humankind, but the end of the Old Covenant between God and Israel; which, according to preterism, took place when the Temple in Jerusalem was destroyed in 70 CE. Preterists believe that prophecies—such as the Second Coming, the desecration of the Jewish Temple, the destruction of Jerusalem, the rise of the Antichrist, the Great Tribulation, the advent of The Day of the Lord, and a Final Judgment—had been fulfilled when the Romans sacked Jerusalem and completely destroyed its Temple. Proponents of full preterism do not believe in a coming resurrection of the dead. They place this event (as well as the Second Coming) in the year 70. Advocates of partial preterism do believe in a coming resurrection. Full preterists contend that partial preterists are merely futurists, since they believe the Second Coming, the Resurrection, the Rapture, and the Judgment are yet to come. Many preterists believe first-century Christians experienced the Rapture to rejoin the Christ. According with Preterism's interpretation of end times, many "time passages" in the New Testament foretell a Second Coming of Christ, with last days to take place within the lifetimes of his disciples: Matt. 10:23, Matt. 16:28, Matt. 24:34, Matt. 26:64, Rom. 13:11–12, 1 Cor. 7:29–31, 1 Cor. 10:11, Phil. 4:5, James 5:8–9, 1 Pet. 4:7, 1 Jn. 2:18. Dispensationalists Dispensationalism is an evangelical futurist Biblical interpretation that foresees a series of dispensations, or periods, in which God relates to human beings under different Biblical covenants. The belief system is primarily rooted in the writings of John Nelson Darby and is premillennial in content. The reestablishment of Israel in 1948 provided a major impetus to the dispensationalist belief system. The wars of Israel after 1948 with its Arab neighbors provided further support, according to John F. Walvoord. After the Six-Day War in 1967, and the Yom Kippur War in 1973, it seemed plausible to many Fundamentalist Christians in the 1970s that Middle East turmoil may well be leading up to the fulfillment of various Bible prophecies and to the Battle of Armageddon. Members of the dispensationalist movement such as Hal Lindsey, J. Dwight Pentecost, John Walvoord, all of whom have Dallas Theological Seminary backgrounds, and some other writers, claimed further that the European Economic Community, which preceded the European Union, would become a United States of Europe, which would in turn become a Revived Roman Empire ruled by the Antichrist. The Revived Roman Empire also figured into the New Testament writers' vision of the future. The fact that in the early 1970s, there were (erroneously thought to be) seven nations in the European Economic Community was held to be significant; this aligned the Community with a seven-headed beast mentioned in Revelation. This specific prophecy has required revision, but the idea of a Revived Roman Empire remains. Dispensationalism, in contrast to the Millerite Adventist movement, had its beginning in the 19th century, when John Nelson Darby, founder of the Plymouth Brethren religious denomination, incorporated into his system of Biblical interpretation a system of organizing Biblical time into a number of discrete dispensations, each of which marks a separate covenant with God. Darby's beliefs were widely publicized in Cyrus I. Scofield's Scofield Reference Bible, an annotated Bible that became popular in the United States. Since the majority of the Biblical prophets were writing at a time when the Temple in Jerusalem was still functioning, they wrote as if it would still be standing during the prophesied events. According to preterism, this was a fulfillment of the prophecies. However, according to Futurists, their destruction in AD 70 put the prophetic timetable on hold. Many such believers therefore anticipated the return of Jews to Israel and the reconstruction of the Temple before the Second Coming could occur. Post-tribulation pre-millennialism A view of the Second Coming of Christ as held by post-tribulational pre-millennialists holds that the Church of Christ will have to undergo great persecution by being present during the great tribulation. Specific prophetic movements In 1843, William Miller made the first of several predictions that the world would end in only a few months. As his predictions did not come true (referred to as the Great Disappointment), followers of Miller went on to found separate groups, the most successful of which is the Seventh-day Adventist Church. Members of the Baháʼí Faith believe Miller's interpretation of signs and dates of the coming of Jesus were, for the most part, correct. They believe the fulfillment of biblical prophecies of the coming of Christ came through a forerunner of their own religion, the Báb. According to the Báb's words, 4 April 1844 was "the first day that the Spirit descended" into his heart. His subsequent declaration to Mullá Husayn-i Bushru'i that he was the "Promised One"—an event now commemorated by Baháʼís as a major holy day—took place on 23 May 1844. It was in October of that year that the Báb embarked on a pilgrimage to Mecca, where he openly declared his claims to the Sharif of Mecca. The first news coverage of these events in the West was in 1845 by The Times, followed by others in 1850 in the United States. The first Baháʼí to come to America was in 1892. Several Baháʼí books and pamphlets make mention of the Millerites, the prophecies used by Miller and the Great Disappointment, most notably William Sears's Thief in the Night. Restorationism (Christian primitivism) End times theology is also significant to restorationist Christian religions, which consider themselves distinct from both Catholicism and Protestantism. Jehovah's Witnesses The eschatology of Jehovah's Witnesses is central to their religious beliefs. They believe Jesus Christ has been ruling in heaven as king since 1914 (a date they believe was prophesied in the Bible) and that after that time a period of cleansing occurred, resulting in God's selection of the Bible Students associated with Charles Taze Russell as his people in 1919. They also believe that the destruction of those who reject the Bible's message and thus willfully refuse to obey God will shortly take place at Armageddon, ensuring that the beginning of the new earthly society will be composed of willing subjects of that kingdom. The religion's doctrines surrounding 1914 are the legacy of a series of emphatic claims regarding the years 1799, 1874, 1878, 1914, 1918 and 1925 made in the Watch Tower Society's publications between 1879 and 1924. Claims about the significance of those years, including the presence of Jesus Christ, the beginning of the "last days", the destruction of worldly governments and the earthly resurrection of Jewish patriarchs, were successively abandoned. In 1922 the society's principal magazine, The Watchtower, described its chronology as "no stronger than its weakest link", but also claimed the chronological relationships to be "of divine origin and divinely corroborated... in a class by itself, absolutely and unqualifiedly correct" and "indisputable facts", and repudiation of Russell's teachings was described as "equivalent to a repudiation of the Lord". The Watch Tower Society has acknowledged its early leaders promoted "incomplete, even inaccurate concepts". The Governing Body of Jehovah's Witnesses says that, unlike Old Testament prophets, its interpretations of the Bible are not inspired or infallible. It says that Bible prophecies can be fully understood only after their fulfillment, citing examples of biblical figures who did not understand the meaning of prophecies they received. Watch Tower Society literature often cites Proverbs 4:18, "The path of the righteous ones is like the bright light that is getting lighter and lighter until the day is firmly established" (NWT) to support their view that there would be an increase in knowledge during "the time of the end", and that this increase in knowledge needs adjustments. Watch Tower Society publications also say that unfulfilled expectations are partly due to eagerness for God's Kingdom and that they do not call their core beliefs into question. The Church of Jesus Christ of Latter-day Saints Members of the Church of Jesus Christ of Latter-day Saints (LDS Church) believe there will be a Second Coming of Jesus to the earth at some time in the future. The LDS Church and its leaders do not make any predictions of the date of the Second Coming. According to church doctrine, the true gospel will be taught in all parts of the world prior to the Second Coming. They also believe there will be increasing war, earthquakes, hurricanes, and man-made disasters prior to the Second Coming. Disasters of all kind will happen before Christ comes. Upon the return of Jesus Christ, all people will be resurrected, the righteous in a first resurrection and the unrighteous in a second, later resurrection. Christ shall reign for a period of 1000 years, after which the Final Judgement will occur. Realized eschatology Realized eschatology is a Christian eschatological theory that holds that the eschatological passages in the New Testament do not refer to the future, but instead refer to the ministry of Jesus and his lasting legacy. Islam Muslims believe there are three periods before the Day of Judgment with some debate as to whether the periods could overlap. Sunni Sunnis believe the dead will then stand in a grand assembly, awaiting a scroll detailing their righteous deeds, sinful acts and ultimate judgment. Prophet Muhammad will be the first to be resurrected. Punishments will include adhab, or severe pain and embarrassment, and khizy or shame. There will also be a punishment of the grave between death and the resurrection. Several Sunni scholars explain some of the signs metaphorically. The signs of the coming end time are divided into major and minor signs: Following the second period, the third is said to be marked by the ten major signs known as alamatu's-sa'ah al- kubra (The major signs of the end). They are as follows: A huge black cloud of smoke (dukhan) will cover the earth. Three sinkings of the earth, one in the East. One sinking of the earth in the West. One sinking of the earth in Arabia. The false messiah—anti-Christ, Masih ad-Dajjal—shall appear with great powers as a one-eyed man with his right eye blind and deformed like a grape. Although believers will not be deceived, he will claim to be God, to hold the keys to heaven and hell, and will lead many astray. In reality, his heaven is hell, and his hell is heaven. The Dajjal will be followed by seventy thousand Jews of Isfahan wearing Persian shawls. The return of Isa (Jesus), from the fourth sky, to kill Dajjal. Ya'jooj and Ma'jooj (Gog and Magog), a Japhetic tribe of vicious beings who had been imprisoned by Dhul-Qarnayn, will break out. They will ravage the earth, drink all the water of Lake Tiberias, and kill all believers in their way. Isa, Imam Al-Mahdi, and the believers with them will go to the top of a mountain and pray for the destruction of Gog and Magog. God eventually will send disease and worms to wipe them out. The sun will rise from the West. The Dabbat al-ard, or Beast of the Earth, will come out of the ground to talk to people. The second blow of the trumpet will be sounded, the dead will return to life, and a fire will come out of Yemen that shall gather all to Mahshar Al Qiy'amah (The Gathering for Judgment). Shia Many of the signs shown above are shared by both Sunni and Shia beliefs, with some exceptions, e.g. Imam Al-Mahdi defeating Al-Masih ad-Dajjal. Concepts and terminology in Shia eschatology include Mi'ad, the Occultation, Al-Yamani, and Sufyani. In Twelver Shia narrations about the last days, the literature largely revolves around Muhammad al-Mahdi, who is considered by many beliefs to be the true twelfth appointed successor to Muhammad. Muhammad al-Mahdi will help mankind against the deception by the Dajjal who will try to get people in to a new world religion which is called "the great deception". Ahmadiyya Ahmadiyya is considered distinct from mainstream Islam. In its writing, the present age has been witness to the evil of man and wrath of God, with war and natural disaster. Ghulam Ahmad is seen as the promised Messiah and the Mahdi, fulfilling Islamic and Biblical prophecies, as well as scriptures of other religions such as Hinduism. His teaching will establish spiritual reform and establish an age of peace. This will continue for a thousand years, and will unify mankind under one faith. Ahmadis believe that despite harsh and strong opposition and discrimination they will eventually be triumphant and their message vindicated both by Muslims and non-Muslims alike. Ahmadis also incorporate the eschatological views from other religions into their doctrine and believe Mirza Ghulam Ahmed falls into this sequence. Baháʼí Faith In the Baháʼí Faith, creation has neither a beginning nor an end; Baháʼís regard the eschatologies of other religions as symbolic. In Baháʼí belief, human time is marked by a series of progressive revelations in which successive messengers or prophets come from God. The coming of each of these messengers is seen as the day of judgment to the adherents of the previous religion, who may choose to accept the new messenger and enter the "heaven" of belief, or denounce the new messenger and enter the "hell" of denial. In this view, the terms "heaven" and "hell" become symbolic terms for a person's spiritual progress and their nearness to or distance from God. In Baháʼí belief, Bahá'u'lláh (1817-1892), the founder of the Baháʼí Faith, was the Second Coming of Christ and also the fulfilment of previous eschatological expectations of Islam and other major religions. The inception of the Baháʼí Faith coincides with Great Disappointment of the Millerite prophesy in 1844. ʻAbdu'l-Bahá taught that Armageddon would begin in 1914, but without a clear indication of its end date. Baháʼís believe that the mass martyrdom anticipated during the End Times had already passed within the historical context of the Baháʼí Faith. Baháʼís expect their faith to be eventually embraced by the masses of the world, ushering in a golden age. Rastafari Rastafari have a unique interpretation of end times, based on the Old Testament and the Book of Revelation. They believe Ethiopian Emperor Haile Selassie I to be God incarnate, the King of kings and Lord of lords mentioned in Revelation 5:5. They saw the crowning of Selassie as the second coming, and the Second Italo-Ethiopian War as fulfillment of Revelation. There is also the expectation that Selassie will return for a day of judgment and bring home the "lost children of Israel", which in Rastafari refers to those taken from Africa through the slave trade. There will then be an era of peace and harmony at Mount Zion in Africa. Cyclic cosmology Hinduism The Vaishnavite tradition links contemporary Hindu eschatology to the figure of Kalki, the tenth and last avatar of Vishnu. Many Hindus believe that before the age draws to a close, Kalki will reincarnate as Shiva and simultaneously dissolve and regenerate the universe. In contrast, Shaivites hold the view that Shiva is incessantly destroying and creating the world. In Hindu eschatology, time is cyclic and consists of kalpas. Each lasts 4.1–8.2 billion years, which is a period of one full day and night for Brahma, who will be alive for 311 trillion, 40 billion years. Within a kalpa there are periods of creation, preservation and decline. After this larger cycle, all of creation will contract to a singularity and then again will expand from that single point, as the ages continue in a religious fractal pattern. Within the current kalpa, there are four epochs that encompass the cycle. They progress from a beginning of complete purity to a descent into total corruption. The last of the four ages is Kali Yuga (which most Hindus believe is the current time), characterized by quarrel, hypocrisy, impiety, violence and decay. The four pillars of dharma will be reduced to one, with truth being all that remains. As written in the Gita: Yadā yadā hi dharmasya glānirbhavati Bhārata Abhyutthānam adharmasya tadātmānam sṛjāmyaham Whenever there is decay of righteousness in Bharata (Aryavarta) And a rise of unrighteousness then I manifest Myself! At this time of chaos, the final avatar, Kalki, endowed with eight superhuman faculties will appear on a white horse. Kalki will amass an army to "establish righteousness upon the earth" and leave "the minds of the people as pure as crystal." At the completion of Kali Yuga, the next Yuga Cycle will begin with a new Satya Yuga, in which all will once again be righteous with the reestablishment of dharma. This, in turn, will be followed by epochs of Treta Yuga, Dvapara Yuga and again another Kali Yuga. This cycle will then repeat until the larger cycle of existence under Brahma returns to the singularity, and a new universe is born. The cycle of birth, growth, decay, and renewal at the individual level finds its echo in the cosmic order, yet is affected by vagueries of divine intervention in Vaishnavite belief. Buddhism There is no classic account of beginning or end in Buddhism; Masao Abe attributes this to the absence of God. History is embedded in the continuing process of samsara or the "beginningless and endless cycles of birth-death-rebirth". Buddhists believe there is an end to things but it is not final because they are bound to be born again. However, the writers of Mahayana Buddhist scriptures establish a specific end-time account in Buddhist tradition: this describes the return of Maitreya Buddha, who would bring about an end to the world. This constitutes one of the two major branches of Buddhist eschatology, with the other being the Sermon of the Seven Suns. End time in Buddhism could also involve a cultural eschatology covering "final things", which include the idea that Sakyamuni Buddha's dharma will also come to an end. Maitreya The Buddha described his teachings disappearing five thousand years from when he preached them, corresponding approximately to the year 4300 since he was born in 623 BCE. At this time, knowledge of dharma will be lost as well. The last of his relics will be gathered in Bodh Gaya and cremated. There will be a new era in which the next Buddha Maitreya will appear, but it will be preceded by the degeneration of human society. This will be a period of greed, lust, poverty, ill will, violence, murder, impiety, physical weakness, sexual depravity and societal collapse, and even the Buddha himself will be forgotten. This will be followed by the coming of Maitreya when the teachings of dharma are forgotten. Maitreya was the first Bodhisattva around whom a cult developed, in approximately the third century CE. The earliest known mention of Maitreya occurs in the Cakavatti, or Sihanada Sutta in Digha Nikaya 26 of the Pali Canon. In it, Gautama Buddha predicted his teachings of dharma would be forgotten after 5,000 years. The text then foretells the birth of Maitreya Buddha in the city of Ketumatī in present-day Benares, whose king will be the Cakkavattī Sankha. Sankha will live in the former palace of King Mahāpanadā, and will become a renunciate who follows Maitreya. In Mahayana Buddhism, Maitreya will attain bodhi in seven days, the minimum period, by virtue of his many lifetimes of preparation. Once Buddha, he will rule over the Ketumati Pure Land, an earthly paradise sometimes associated with the Indian city of Varanasi or Benares in present-day Uttar Pradesh. In Mahayana Buddhism, the Buddha presides over a land of purity. For example, Amitabha presides over Sukhavati, more popularly known as the "Western Paradise". A notable teaching he will rediscover is that of the ten non-virtuous deeds—killing, stealing, sexual misconduct, lying, divisive speech, abusive speech, idle speech, covetousness, harmful intent and wrong views. The ten virtuous deeds will replace them with the abandonment of each of these practices. Edward Conze in his Buddhist Scriptures (1959) gives an account of Maitreya: Maitreya currently resides in Tushita, but will come to Jambudvipa when needed most as successor to the historic Śākyamuni Buddha. Maitreya will achieve complete enlightenment during his lifetime, and following this reawakening he will bring back the timeless teaching of dharma to this plane and rediscover enlightenment. The Arya Maitreya Mandala, founded in 1933 by Lama Anagarika Govinda, is based on the idea of Maitreya. Maitreya eschatology forms the central canon of the White Lotus Society, a religious and political movement which emerged in Yuan China. It later branched into the Chinese underground criminal organization known as the Triads, which exist today as an international underground criminal network. Note that no description of Maitreya occurs in any other sutta in the canon, casting doubt as to the authenticity of the scripture. In addition, sermons of the Buddha normally are in response to a question, or in a specific context, but this sutta has a beginning and an ending, and its content is quite different from the others. This has led some to conclude that the whole sutta is apocryphal, or tampered with. Sermon of the Seven Suns In his "Sermon of the Seven Suns" in the Pali Canon, the Buddha describes the ultimate fate of the Earth in an apocalypse characterized by the consequent appearance of seven suns in the sky, each causing progressive ruin until the planet is destroyed: The canon goes on to describe the progressive destruction of each sun. The third sun will dry the Ganges River and other rivers, whilst the fourth will cause the lakes to evaporate; the fifth will dry the oceans. Later: The sermon completes with the Earth immersed into an extensive holocaust. The Pali Canon does not indicate when this will happen relative to Maitreya. Norse mythology Norse mythology depicts the end of days as Ragnarök, an Old Norse term translatable as "twilight of the gods". It will be heralded by a devastation known as Fimbulvetr which will seize Midgard in cold and darkness. The sun and moon will disappear from the sky, and poison will fill the air. The dead will rise from the ground and there will be widespread despair. There follows a battle between—on the one hand—the Gods with the Æsir, Vanir and Einherjar, led by Odin, and—on the other hand—forces of Chaos, including the fire giants and jötunn, led by Loki. In the fighting Odin will be swallowed whole by his old nemesis Fenrir. The god Freyr fights Surtr but loses. Víðarr, son of Odin, will then avenge his father by ripping Fenrir's jaws apart and stabbing the wolf in the heart with his spear. The serpent Jörmungandr will open its gaping maw and be met in combat by Thor. Thor, also a son of Odin, will defeat the serpent, only to take nine steps afterwards before collapsing in his own death. After this people will flee their homes as the sun blackens and the earth sinks into the sea. The stars will vanish, steam will rise, and flames will touch the heavens. This conflict will result in the deaths of most of the major Gods and forces of Chaos. Finally, Surtr will fling fire across the nine worlds. The ocean will then completely submerge Midgard. After the cataclysm, the world will resurface new and fertile, and the surviving Gods will meet. Baldr, another son of Odin, will be reborn in the new world, according to Völuspá. The two human survivors, Líf and Lífþrasir, will then repopulate this new earth. No end times Taoism The Taoist faith is not concerned with what came before or after life, knowing only their own being in the Tao. The philosophy is that people come and go, just like mountains, trees and stars, but Tao will go on for time immemorial. Analogies in science and philosophy Researchers in futures studies and transhumanists investigate how the accelerating rate of scientific progress may lead to a "technological singularity" in the future that would profoundly and unpredictably change the course of human history, and result in Homo sapiens no longer being the dominant life form on Earth. Occasionally the term "physical eschatology" is applied to the long-term predictions of astrophysics about the future of Earth and ultimate fate of the universe. The Sun will turn into a red giant in approximately 6 billion years. Life on Earth will become impossible due to a rise in temperature long before the planet is actually swallowed up by the Sun. Even later, the Sun will become a white dwarf. See also Notes References Further reading Craig C. Hill, In God's Time: The Bible and the Future, Grand Rapids: Eerdmans 2002. . Dave Hunt, A Cup of Trembling, Harvest House Publishers, Eugene, (Oregon) 1995 . Jonathan Menn, Biblical Eschatology, Eugene, Oregon, Wipf & Stock 2013. . Joseph Ratzinger., Eschatology: Death and Eternal Life, Washington D.C.: Catholic University of America Press 1985. . Robert Sungenis, Scott Temple, David Allen Lewis, Shock Wave 2000! subtitled The Harold Camping 1994 Debacle, New Leaf Press, Inc. 2004, . Stephen Travis, Christ Will Come Again: Hope for the Second Coming of Jesus, Toronto: Clements Publishing 2004. . Jerry L. Walls (ed.), The Oxford Handbook of Eschatology, New York: Oxford University Press 2008. . External links Christian Eschatology Explained (A short primer on the main schools of Christian Eschatology) Prophecy Mythology Last events Ultimate fate of the universe
9927
https://en.wikipedia.org/wiki/Endomembrane%20system
Endomembrane system
The endomembrane system is composed of the different membranes (endomembranes) that are suspended in the cytoplasm within a eukaryotic cell. These membranes divide the cell into functional and structural compartments, or organelles. In eukaryotes the organelles of the endomembrane system include: the nuclear membrane, the endoplasmic reticulum, the Golgi apparatus, lysosomes, vesicles, endosomes, and plasma (cell) membrane among others. The system is defined more accurately as the set of membranes that forms a single functional and developmental unit, either being connected directly, or exchanging material through vesicle transport. Importantly, the endomembrane system does not include the membranes of plastids or mitochondria, but might have evolved partially from the actions of the latter (see below). The nuclear membrane contains a lipid bilayer that encompasses the contents of the nucleus. The endoplasmic reticulum (ER) is a synthesis, and transport organelle that branches into the cytoplasm in plant and animal cells. The Golgi apparatus is a series of multiple compartments where molecules are packaged for delivery to other cell components or for secretion from the cell. Vacuoles, which are found in both plant and animal cells (though much bigger in plant cells), are responsible for maintaining the shape and structure of the cell as well as storing waste products. A vesicle is a relatively small, membrane-enclosed sac that stores or transports substances. The cell membrane is a protective barrier that regulates what enters and leaves the cell. There is also an organelle known as the Spitzenkörper that is only found in fungi, and is connected with hyphal tip growth. In prokaryotes endomembranes are rare, although in many photosynthetic bacteria the plasma membrane is highly folded and most of the cell cytoplasm is filled with layers of light-gathering membrane. These light-gathering membranes may even form enclosed structures called chlorosomes in green sulfur bacteria. Another example is the complex "pepin" system of Thiomargarita species, especially T. magnifica. The organelles of the endomembrane system are related through direct contact or by the transfer of membrane segments as vesicles. Despite these relationships, the various membranes are not identical in structure and function. The thickness, molecular composition, and metabolic behavior of a membrane are not fixed, they may be modified several times during the membrane's life. One unifying characteristic the membranes share is a lipid bilayer, with proteins attached to either side or traversing them. History of the concept Most lipids are synthesized in yeast either in the endoplasmic reticulum, lipid particles, or the mitochondrion, with little or no lipid synthesis occurring in the plasma membrane or nuclear membrane. Sphingolipid biosynthesis begins in the endoplasmic reticulum, but is completed in the Golgi apparatus. The situation is similar in mammals, with the exception of the first few steps in ether lipid biosynthesis, which occur in peroxisomes. The various membranes that enclose the other subcellular organelles must therefore be constructed by transfer of lipids from these sites of synthesis. However, although it is clear that lipid transport is a central process in organelle biogenesis, the mechanisms by which lipids are transported through cells remain poorly understood. The first proposal that the membranes within cells form a single system that exchanges material between its components was by Morré and Mollenhauer in 1974. This proposal was made as a way of explaining how the various lipid membranes are assembled in the cell, with these membranes being assembled through lipid flow from the sites of lipid synthesis. The idea of lipid flow through a continuous system of membranes and vesicles was an alternative to the various membranes being independent entities that are formed from transport of free lipid components, such as fatty acids and sterols, through the cytosol. Importantly, the transport of lipids through the cytosol and lipid flow through a continuous endomembrane system are not mutually exclusive processes and both may occur in cells. Components of the system Nuclear envelope The nuclear envelope surrounds the nucleus, separating its contents from the cytoplasm. It has two membranes, each a lipid bilayer with associated proteins. The outer nuclear membrane is continuous with the rough endoplasmic reticulum membrane, and like that structure, features ribosomes attached to the surface. The outer membrane is also continuous with the inner nuclear membrane since the two layers are fused together at numerous tiny holes called nuclear pores that perforate the nuclear envelope. These pores are about 120 nm in diameter and regulate the passage of molecules between the nucleus and cytoplasm, permitting some to pass through the membrane, but not others. Since the nuclear pores are located in an area of high traffic, they play an important role in cell physiology. The space between the outer and inner membranes is called the perinuclear space and is joined with the lumen of the rough ER. The nuclear envelope's structure is determined by a network of intermediate filaments (protein filaments). This network is organized into lining similar to mesh called the nuclear lamina, which binds to chromatin, integral membrane proteins, and other nuclear components along the inner surface of the nucleus. The nuclear lamina is thought to help materials inside the nucleus reach the nuclear pores and in the disintegration of the nuclear envelope during mitosis and its reassembly at the end of the process. The nuclear pores are highly efficient at selectively allowing the passage of materials to and from the nucleus, because the nuclear envelope has a considerable amount of traffic. RNA and ribosomal subunits must be continually transferred from the nucleus to the cytoplasm. Histones, gene regulatory proteins, DNA and RNA polymerases, and other substances essential for nuclear activities must be imported from the cytoplasm. The nuclear envelope of a typical mammalian cell contains 3000–4000 pore complexes. If the cell is synthesizing DNA each pore complex needs to transport about 100 histone molecules per minute. If the cell is growing rapidly, each complex also needs to transport about 6 newly assembled large and small ribosomal subunits per minute from the nucleus to the cytosol, where they are used to synthesize proteins. Endoplasmic reticulum The endoplasmic reticulum (ER) is a membranous synthesis and transport organelle that is an extension of the nuclear envelope. More than half the total membrane in eukaryotic cells is accounted for by the ER. The ER is made up of flattened sacs and branching tubules that are thought to interconnect, so that the ER membrane forms a continuous sheet enclosing a single internal space. This highly convoluted space is called the ER lumen and is also referred to as the ER cisternal space. The lumen takes up about ten percent of the entire cell volume. The endoplasmic reticulum membrane allows molecules to be selectively transferred between the lumen and the cytoplasm, and since it is connected to the nuclear envelope, it provides a channel between the nucleus and the cytoplasm. The ER has a central role in producing, processing, and transporting biochemical compounds for use inside and outside of the cell. Its membrane is the site of production of all the transmembrane proteins and lipids for most of the cell's organelles, including the ER itself, the Golgi apparatus, lysosomes, endosomes, mitochondria, peroxisomes, secretory vesicles, and the plasma membrane. Furthermore, almost all of the proteins that will exit the cell, plus those destined for the lumen of the ER, Golgi apparatus, or lysosomes, are originally delivered to the ER lumen. Consequently, many of the proteins found in the cisternal space of the endoplasmic reticulum lumen are there only temporarily as they pass on their way to other locations. Other proteins, however, constantly remain in the lumen and are known as endoplasmic reticulum resident proteins. These special proteins contain a specialized retention signal made up of a specific sequence of amino acids that enables them to be retained by the organelle. An example of an important endoplasmic reticulum resident protein is the chaperone protein known as BiP which identifies other proteins that have been improperly built or processed and keeps them from being sent to their final destinations. The ER is involved in cotranslational sorting of proteins. A polypeptide which contains an ER signal sequence is recognised by the signal recognition particle which halts the production of the protein. The SRP transports the nascent protein to the ER membrane where it is released through a membrane channel and translation resumes. There are two distinct, though connected, regions of ER that differ in structure and function: smooth ER and rough ER. The rough endoplasmic reticulum is so named because the cytoplasmic surface is covered with ribosomes, giving it a bumpy appearance when viewed through an electron microscope. The smooth ER appears smooth since its cytoplasmic surface lacks ribosomes. Functions of the smooth ER In the great majority of cells, smooth ER regions are scarce and are often partly smooth and partly rough. They are sometimes called transitional ER because they contain ER exit sites from which transport vesicles carrying newly synthesized proteins and lipids bud off for transport to the Golgi apparatus. In certain specialized cells, however, the smooth ER is abundant and has additional functions. The smooth ER of these specialized cells functions in diverse metabolic processes, including synthesis of lipids, metabolism of carbohydrates, and detoxification of drugs and poisons. Enzymes of the smooth ER are vital to the synthesis of lipids, including oils, phospholipids, and steroids. Sex hormones of vertebrates and the steroid hormones secreted by the adrenal glands are among the steroids produced by the smooth ER in animal cells. The cells that synthesize these hormones are rich in smooth ER. Liver cells are another example of specialized cells that contain an abundance of smooth ER. These cells provide an example of the role of smooth ER in carbohydrate metabolism. Liver cells store carbohydrates in the form of glycogen. The breakdown of glycogen eventually leads to the release of glucose from the liver cells, which is important in the regulation of sugar concentration in the blood. However, the primary product of glycogen breakdown is glucose-1-phosphate. This is converted to glucose-6-phosphate and then an enzyme of the liver cell's smooth ER removes the phosphate from the glucose, so that it can then leave the cell. Enzymes of the smooth ER can also help detoxify drugs and poisons. Detoxification usually involves the addition of a hydroxyl group to a drug, making the drug more soluble and thus easier to purge from the body. One extensively studied detoxification reaction is carried out by the cytochrome P450 family of enzymes, which catalyze water-insoluble drugs or metabolites that would otherwise accumulate to toxic levels in cell membrane. Muscle cells have another specialized function of smooth ER. The ER membrane pumps calcium ions from the cytosol into the cisternal space. When a muscle cell becomes stimulated by a nerve impulse, calcium goes back across the ER membrane into the cytosol and generates the contraction of the muscle cell. Functions of the rough ER Many types of cells export proteins produced by ribosomes attached to the rough ER. The ribosomes assemble amino acids into protein units, which are carried into the rough ER for further adjustments. These proteins may be either transmembrane proteins, which become embedded in the membrane of the endoplasmic reticulum, or water-soluble proteins, which are able to pass through the membrane into the lumen. Those that reach the inside of the endoplasmic reticulum are folded into the correct three-dimensional conformation. Chemicals, such as carbohydrates or sugars, are added, then the endoplasmic reticulum either transports the completed proteins, called secretory proteins, to areas of the cell where they are needed, or they are sent to the Golgi apparatus for further processing and modification. Once secretory proteins are formed, the ER membrane separates them from the proteins that will remain in the cytosol. Secretory proteins depart from the ER enfolded in the membranes of vesicles that bud like bubbles from the transitional ER. These vesicles in transit to another part of the cell are called transport vesicles. An alternative mechanism for transport of lipids and proteins out of the ER are through lipid transfer proteins at regions called membrane contact sites where the ER becomes closely and stably associated with the membranes of other organelles, such as the plasma membrane, Golgi or lysosomes. In addition to making secretory proteins, the rough ER makes membranes that grows in place from the addition of proteins and phospholipids. As polypeptides intended to be membrane proteins grow from the ribosomes, they are inserted into the ER membrane itself and are kept there by their hydrophobic portions. The rough ER also produces its own membrane phospholipids; enzymes built into the ER membrane assemble phospholipids. The ER membrane expands and can be transferred by transport vesicles to other components of the endomembrane system. Golgi apparatus The Golgi apparatus (also known as the Golgi body and the Golgi complex) is composed of separate sacs called cisternae. Its shape is similar to a stack of pancakes. The number of these stacks varies with the specific function of the cell. The Golgi apparatus is used by the cell for further protein modification. The section of the Golgi apparatus that receives the vesicles from the ER is known as the cis face, and is usually near the ER. The opposite end of the Golgi apparatus is called the trans face, this is where the modified compounds leave. The trans face is usually facing the plasma membrane, which is where most of the substances the Golgi apparatus modifies are sent. Vesicles sent off by the ER containing proteins are further altered at the Golgi apparatus and then prepared for secretion from the cell or transport to other parts of the cell. Various things can happen to the proteins on their journey through the enzyme covered space of the Golgi apparatus. The modification and synthesis of the carbohydrate portions of glycoproteins is common in protein processing. The Golgi apparatus removes and substitutes sugar monomers, producing a large variety of oligosaccharides. In addition to modifying proteins, the Golgi also manufactures macromolecules itself. In plant cells, the Golgi produces pectins and other polysaccharides needed by the plant structure. Once the modification process is completed, the Golgi apparatus sorts the products of its processing and sends them to various parts of the cell. Molecular identification labels or tags are added by the Golgi enzymes to help with this. After everything is organized, the Golgi apparatus sends off its products by budding vesicles from its trans face. Vacuoles Vacuoles, like vesicles, are membrane-bound sacs within the cell. They are larger than vesicles and their specific function varies. The operations of vacuoles are different for plant and animal vacuoles. In plant cells, vacuoles cover anywhere from 30% to 90% of the total cell volume. Most mature plant cells contain one large central vacuole encompassed by a membrane called the tonoplast. Vacuoles of plant cells act as storage compartments for the nutrients and waste of a cell. The solution that these molecules are stored in is called the cell sap. Pigments that color the cell are sometime located in the cell sap. Vacuoles can also increase the size of the cell, which elongates as water is added, and they control the turgor pressure (the osmotic pressure that keeps the cell wall from caving in). Like lysosomes of animal cells, vacuoles have an acidic pH and contain hydrolytic enzymes. The pH of vacuoles enables them to perform homeostatic procedures in the cell. For example, when the pH in the cells environment drops, the H+ ions surging into the cytosol can be transferred to a vacuole in order to keep the cytosol's pH constant. In animals, vacuoles serve in exocytosis and endocytosis processes. Endocytosis refers to when substances are taken into the cell, whereas for exocytosis substances are moved from the cell into the extracellular space. Material to be taken-in is surrounded by the plasma membrane, and then transferred to a vacuole. There are two types of endocytosis, phagocytosis (cell eating) and pinocytosis (cell drinking). In phagocytosis, cells engulf large particles such as bacteria. Pinocytosis is the same process, except the substances being ingested are in the fluid form. Vesicles Vesicles are small membrane-enclosed transport units that can transfer molecules between different compartments. Most vesicles transfer the membranes assembled in the endoplasmic reticulum to the Golgi apparatus, and then from the Golgi apparatus to various locations. There are various types of vesicles each with a different protein configuration. Most are formed from specific regions of membranes. When a vesicle buds off from a membrane it contains specific proteins on its cytosolic surface. Each membrane a vesicle travels to contains a marker on its cytosolic surface. This marker corresponds with the proteins on the vesicle traveling to the membrane. Once the vesicle finds the membrane, they fuse. There are three well known types of vesicles. They are clathrin-coated, COPI-coated, and COPII-coated vesicles. Each performs different functions in the cell. For example, clathrin-coated vesicles transport substances between the Golgi apparatus and the plasma membrane. COPI- and COPII-coated vesicles are frequently used for transportation between the ER and the Golgi apparatus. Lysosomes Lysosomes are organelles that contain hydrolytic enzymes that are used for intracellular digestion. The main functions of a lysosome are to process molecules taken in by the cell and to recycle worn out cell parts. The enzymes inside of lysosomes are acid hydrolases which require an acidic environment for optimal performance. Lysosomes provide such an environment by maintaining a pH of 5.0 inside of the organelle. If a lysosome were to rupture, the enzymes released would not be very active because of the cytosol's neutral pH. However, if numerous lysosomes leaked the cell could be destroyed from autodigestion. Lysosomes carry out intracellular digestion, in a process called phagocytosis (from the Greek , to eat and , vessel, referring here to the cell), by fusing with a vacuole and releasing their enzymes into the vacuole. Through this process, sugars, amino acids, and other monomers pass into the cytosol and become nutrients for the cell. Lysosomes also use their hydrolytic enzymes to recycle the cell's obsolete organelles in a process called autophagy. The lysosome engulfs another organelle and uses its enzymes to take apart the ingested material. The resulting organic monomers are then returned to the cytosol for reuse. The last function of a lysosome is to digest the cell itself through autolysis. Spitzenkörper The spitzenkörper is a component of the endomembrane system found only in fungi, and is associated with hyphal tip growth. It is a phase-dark body that is composed of an aggregation of membrane-bound vesicles containing cell wall components, serving as a point of assemblage and release of such components intermediate between the Golgi and the cell membrane. The spitzenkörper is motile and generates new hyphal tip growth as it moves forward. Plasma membrane The plasma membrane is a phospholipid bilayer membrane that separates the cell from its environment and regulates the transport of molecules and signals into and out of the cell. Embedded in the membrane are proteins that perform the functions of the plasma membrane. The plasma membrane is not a fixed or rigid structure, the molecules that compose the membrane are capable of lateral movement. This movement and the multiple components of the membrane are why it is referred to as a fluid mosaic. Smaller molecules such as carbon dioxide, water, and oxygen can pass through the plasma membrane freely by diffusion or osmosis. Larger molecules needed by the cell are assisted by proteins through active transport. The plasma membrane of a cell has multiple functions. These include transporting nutrients into the cell, allowing waste to leave, preventing materials from entering the cell, averting needed materials from leaving the cell, maintaining the pH of the cytosol, and preserving the osmotic pressure of the cytosol. Transport proteins which allow some materials to pass through but not others are used for these functions. These proteins use ATP hydrolysis to pump materials against their concentration gradients. In addition to these universal functions, the plasma membrane has a more specific role in multicellular organisms. Glycoproteins on the membrane assist the cell in recognizing other cells, in order to exchange metabolites and form tissues. Other proteins on the plasma membrane allow attachment to the cytoskeleton and extracellular matrix; a function that maintains cell shape and fixes the location of membrane proteins. Enzymes that catalyze reactions are also found on the plasma membrane. Receptor proteins on the membrane have a shape that matches with a chemical messenger, resulting in various cellular responses. Evolution The origin of the endomembrane system is linked to the origin of eukaryotes themselves and the origin of eukaryoties to the endosymbiotic origin of mitochondria. Many models have been put forward to explain the origin of the endomembrane system (reviewed in). The most recent concept suggests that the endomembrane system evolved from outer membrane vesicles the endosymbiotic mitochondrion secreted, and got enclosed within infoldings of the host prokaryote (in turn, a result of the ingestion of the endosymbiont). This OMV (outer membrane vesicles)-based model for the origin of the endomembrane system is currently the one that requires the fewest novel inventions at eukaryote origin and explains the many connections of mitochondria with other compartments of the cell. Currently, this "inside-out" hypothesis (which states that the alphaproteobacteria, the ancestral mitochondria, were engulfed by the blebs of an asgardarchaeon, and later the blebs fused leaving infoldings which would eventually become the endomembrane system) is favored more than the outside-in one (which suggested that the endomembrane system arose due to infoldings within the archaeal membrane). References Cell anatomy Membrane biology
10078
https://en.wikipedia.org/wiki/Enjambment
Enjambment
In poetry, enjambment ( or ; from the French enjamber) is incomplete syntax at the end of a line; the meaning 'runs over' or 'steps over' from one poetic line to the next, without punctuation. Lines without enjambment are end-stopped. The origin of the word is credited to the French word enjamber, which means 'to straddle or encroach'. In reading, the delay of meaning creates a tension that is released when the word or phrase that completes the syntax is encountered (called the rejet); the tension arises from the "mixed message" produced both by the pause of the line-end, and the suggestion to continue provided by the incomplete meaning. In spite of the apparent contradiction between rhyme, which heightens closure, and enjambment, which delays it, the technique is compatible with rhymed verse. Even in couplets, the closed or heroic couplet was a late development; older is the open couplet, where rhyme and enjambed lines co-exist. Enjambment has a long history in poetry. Homer used the technique, and it is the norm for alliterative verse where rhyme is unknown. In the 32nd Psalm of the Hebrew Bible enjambment is unusually conspicuous. It was used extensively in England by Elizabethan poets for dramatic and narrative verses, before giving way to closed couplets. The example of John Milton in Paradise Lost laid the foundation for its subsequent use by the English Romantic poets; in its preface he identified it as one of the chief features of his verse: "sense variously drawn out from one verse into another". Examples The start of The Waste Land by T. S. Eliot, with only lines 4 and 7 end-stopped: These lines from Shakespeare's The Winter's Tale (c. 1611) are heavily enjambed (meaning enjambment is used): Meaning flows as the lines progress, and the reader's eye is forced to go on to the next sentence. It can also make the reader feel uncomfortable or the poem feel like "flow-of-thought" with a sensation of urgency or disorder. In contrast, the following lines from Romeo and Juliet (c. 1595) are completely end-stopped: Each line is formally correspondent with a unit of thought—in this case, a clause of a sentence. End-stopping is more frequent in early Shakespeare: as his style developed, the proportion of enjambment in his plays increased. Scholars such as Goswin König and A. C. Bradley have estimated approximate dates of undated works of Shakespeare by studying the frequency of enjambment. Endymion by John Keats, lines 2–4: The song "One Night In Bangkok", from the musical Chess, written by Tim Rice and Björn Ulvaeus, includes examples such as : Closely related to enjambment is the technique of "broken rhyme" or "split rhyme" which involves the splitting of an individual word, typically to allow a rhyme with one or more syllables of the split word. In English verse, broken rhyme is used almost exclusively in light verse, such as to form a word that rhymes with "orange", as in this example by Willard Espy, in his poem "The Unrhymable Word: Orange": The clapping game "Miss Susie" uses the break "... Hell / -o operator" to allude to the taboo word "Hell", then replaces it with the innocuous "Hello". See also Blank verse Caesura Concrete poetry Free verse Line break (poetry) Notes References Further reading John Hollander, Vision and Resonance, Oxford U. Press, 1975 (especially chapter 5). Free online explanation with examples Poetic devices Poetic rhythm
10248
https://en.wikipedia.org/wiki/Extrapyramidal
Extrapyramidal
Extrapyramidal can refer to: Extrapyramidal system Extrapyramidal symptoms
10400
https://en.wikipedia.org/wiki/History%20of%20Esperanto
History of Esperanto
L. L. Zamenhof developed Esperanto in the 1870s and '80s. Unua Libro, the first print discussion of the language, appeared in 1887. The number of Esperanto speakers have increased gradually since then, without much support from governments and international organizations. Its use has, in some instances, been outlawed or otherwise suppressed. Standardized Yiddish Around 1880, while in Moscow and approximately simultaneously with working on Esperanto, Zamenhof made an aborted attempt to standardize Yiddish, based on his native Bialystok (Northeastern) dialect, as a unifying language for the Jews of the Russian Empire. He even used a Latin alphabet, with the letters ć, h́, ś, ź (the same as in early drafts of Esperanto, later ĉ, ĥ, ŝ, ĵ) and ě for schwa. However, he concluded there was no future for such a project, and abandoned it, dedicating himself to Esperanto as a unifying language for all humankind. Paul Wexler proposed that Esperanto was not an arbitrary pastiche of major European languages but a Latinate relexification of Yiddish, a native language of its founder. This model is generally unsupported by mainstream linguists. Development of the language before publication Zamenhof would later say that he had dreamed of a world language since he was a child. At first he considered a revival of Latin, but after learning it in school he decided it was too complicated to be a common means of international communication. When he learned English, he realised that verb conjugations were unnecessary, and that grammatical systems could be much simpler than he had expected. He still had the problem of memorising a large vocabulary, until he noticed two Russian signs labelled Швейцарская (švejtsarskaja, a porter's lodge – from швейцар švejtsar, a porter) and Кондитерская (konditerskaja, a confectioner's shop – from кондитер konditer, a confectioner). He then realised that a judicious use of affixes could greatly decrease the number of root words needed for communication. He chose to take his vocabulary from Romance and Germanic, the languages that were most widely taught in schools around the world and would therefore be recognisable to the largest number of people. Zamenhof taught an early version of the language to his high-school classmates. Then, for several years, he worked on translations and poetry to refine his creation. In 1895 he wrote, "I worked for six years perfecting and testing the language, even though it had seemed to me in 1878 that it was already completely ready." When he was ready to publish, the Czarist censors would not allow it. Stymied, he spent his time in translating works such as the Bible and Shakespeare. This enforced delay led to continued improvement. In July 1887 he published his Unua Libro (First Book), a basic introduction to the language. This was essentially the language spoken today. Unua Libro to Declaration of Boulogne (1887–1905) Unua Libro was published in 1887. At first the movement grew most in the Russian empire and eastern Europe, but soon spread to western Europe and beyond: to Argentina in 1889; to Canada in 1901; to Algeria, Chile, Japan, Mexico, and Peru in 1903; to Tunisia in 1904; and to Australia, the United States, Guinea, Indochina, New Zealand, Tonkin, and Uruguay in 1905. In its first years Esperanto was used mainly in publications by Zamenhof and early adopters like Antoni Grabowski, in extensive correspondence (mostly now lost), in the magazine La Esperantisto, published from 1889 to 1895 and only occasionally in personal encounters. In 1894, under pressure from Wilhelm Trompeter, the publisher of the magazine La Esperantisto, and some other leading users, Zamenhof reluctantly put forward a radical reform to be voted on by readers. He proposed the reduction of the alphabet to 22 letters (by eliminating the accented letters and most of their sounds), the change of the plural to -i, the use of a positional accusative instead of the ending -n, the removal of the distinction between adjectives and adverbs, the reduction of the number of participles from six to two, and the replacement of the table of correlatives with more Latinate words or phrases. These reforms were overwhelmingly rejected, but some were picked up in subsequent reforms (such as Ido) and criticisms of the language. In the following decade Esperanto spread into western Europe, especially France. By 1905 there were already 27 magazines being published (Auld 1988). A small international conference was held in 1904, leading to the first world congress in August 1905 in Boulogne-sur-Mer, France. There were 688 Esperanto speakers present from 20 nationalities. At this congress, Zamenhof officially resigned his leadership of the Esperanto movement, as he did not want personal prejudice against himself (or anti-Semitism) to hinder the progress of the language. He proposed a declaration on founding principles of the Esperanto movement, which the attendees of the congress endorsed. Declaration of Boulogne to present (1905–present) World congresses have been held every year since 1905, except during the two World Wars. The autonomous territory of Neutral Moresnet, between Belgium and Germany, had a sizable proportion of Esperanto-speakers among its small and multiethnic population. There was a proposal to make Esperanto its official language. In 1908, it was eventually accepted alongside Dutch, German, and French. In the early 1920s, a great opportunity seemed to arise for Esperanto when the Iranian delegation to the League of Nations proposed that it be adopted for use in international relations, following a report by Nitobe Inazō, an official delegate of League of Nations during the 13th World Congress of Esperanto in Prague. Ten delegates accepted the proposal with only one voice against, the French delegate, Gabriel Hanotaux, who employed France's Council veto privilege to squash all League attempts at the recognition of Esperanto, starting on the first vote on 18 December 1920 and continuing through the next three years. Hanotaux did not like how the French language was losing its position as the international language and saw Esperanto as a threat. However, two years later the League recommended that its member states include Esperanto in their educational curricula. The French retaliated by banning all instruction in Esperanto in French schools and universities. The French Ministry of Instruction said that acceptance of Esperanto would mean that "French and English would perish and that the literary standard of the world would be debased". Nonetheless, many people see the 1920s as the heyday of the Esperanto movement. In 1941, the Soviet Union started performing mass arrests, deportations, and killings of many Esperantists and their relatives for fear of an anti-nationalistic movement, but it was interrupted by the Nazi invasion. Hitler wrote in Mein Kampf that Esperanto was created as a universal language to unite the Jewish diaspora. The creation of a Jew-free National German Esperanto League was not enough to placate the Nazis. The teaching of Esperanto was not allowed in German prisoner-of-war camps during World War II. Esperantists sometimes were able to get around the ban by convincing guards that they were teaching Italian, the language of Germany's closest ally. In the early years of the Soviet Union, Esperanto was given a measure of government support, and an officially recognized Soviet Esperanto Association came into being. However, in 1937, Stalin reversed this policy and the use of Esperanto was effectively banned until 1956. While Esperanto itself was not enough cause for execution, its use was extended among Jews or trade unionists and encouraged contacts with foreigners. Fascist Italy, on the other hand, made some efforts of promoting tourism in Italy through Esperanto leaflets and appreciated the similarities of Italian and Esperanto. Portugal's right-wing governments cracked down on the language from 1936 until the Carnation Revolution of 1974. After the Spanish Civil War, Francoist Spain cracked down on the Anarchists and Catalan nationalists among whom the speaking of Esperanto had been quite widespread; but in the 1950s, the Esperanto movement was tolerated again, with Francisco Franco accepting the honorary patronage of the Madrid World Esperanto Congress. The Cold War, especially in the 1950s and 1960s, put a damper on the Esperanto movement as well, as there were fears on both sides that Esperanto could be used for enemy propaganda. However, the language experienced something of a renaissance in the 1970s and spread to new parts of the world, such as its veritable explosion in popularity in Iran in 1975. By 1991 there were enough African Esperantists to warrant a pan-African congress. The language continues to spread, although it is not officially recognised by any country, and is part of the state educational curriculum of only a few. Evolution of the language The Declaration of Boulogne (1905) limited changes to Esperanto. That declaration stated, among other things, that the basis of the language should remain the Fundamento de Esperanto ("Foundation of Esperanto", a group of early works by Zamenhof), which is to be binding forever: nobody has the right to make changes to it. The declaration also permits new concepts to be expressed as the speaker sees fit, but it recommends doing so in accordance with the original style. Many Esperantists believe this declaration stabilising the language is a major reason why the Esperanto speaker community grew beyond the levels attained by other constructed languages and has developed a flourishing culture. Other constructed languages have been hindered from developing a stable speaking community by continual tinkering. Also, many developers of constructed languages have been possessive of their creation and have worked to prevent others from contributing to the language. One such ultimately disastrous case was Schleyer's Volapük. In contrast, Zamenhof declared that "Esperanto belongs to the Esperantists", and moved to the background once the language was published, allowing others to share in the early development of the language. The grammatical description in the earliest books was somewhat vague, so a consensus on usage (influenced by Zamenhof's answers to some questions) developed over time within boundaries set by the initial outline (Auld 1988). Even before the Declaration of Boulogne, the language was remarkably stable; only one set of lexical changes were made in the first year after publication, namely changing "when", "then", "never", "sometimes", "always" from kian, tian, nenian, ian, ĉian to kiam, tiam, neniam etc., to avoid confusion with the accusative forms of kia "what sort of", tia "that sort of", etc. Thus Esperanto achieved a stability of structure and grammar similar to that which natural languages enjoy by virtue of their native speakers and established bodies of literature. One could learn Esperanto without having it move from underfoot. Changes could and did occur in the language, but only by acquiring widespread popular support; there was no central authority making arbitrary changes, as happened with Volapük and some other languages. Modern Esperanto usage may in fact depart from that originally described in the Fundamento, though the differences are largely semantic (involving changed meaning of words) rather than grammatical or phonological. The translation given for "I like this one", in the sample phrases in the main Esperanto article, offers a significant example. According to the Fundamento, Mi ŝatas ĉi tiun would in fact have meant "I esteem this one". The traditional usage is Tiu ĉi plaĉas al mi (literally, "this one is pleasing to me"), which reflects the phrasing of most European languages (French celui-ci me plaît, Spanish éste me gusta, Russian это мне нравится [eto mnye nravitsya], German Das gefällt mir, Italian mi piace). However, the original Ĉi tiu plaĉas al mi continues to be commonly used. For later changes to the language, see Modern evolution of Esperanto. Dialects, reform projects and derived languages Esperanto has not fragmented into regional dialects through natural language use. This may be because it is the language of daily communication for only a small minority of its speakers. However at least three other factors work against dialects, namely the centripetal force of the Fundamento, the unifying influence of the Plena Vortaro and its successors, which exemplified usage from the works of Zamenhof and leading writers, and the transnational ambitions of the speech community itself. Slang and jargon have developed to some extent, but such features interfere with universal communication – the whole point of Esperanto – and so have generally been avoided. However, in the early twentieth century numerous reform projects were proposed. Almost all of these Esperantidos were stillborn, but the very first, Ido ("offspring"), had significant success for several years. Ido was proposed by the Delegation for the Adoption of an International Auxiliary Language in Paris in October 1907. Its main reforms were in bringing the alphabet, semantics, and some grammatical features into closer alignment with the Romance languages, as well as removal of adjectival agreement and the accusative case except when necessary. At first, a number of leading Esperantists put their support behind the Ido project, but the movement stagnated and declined, first with the accidental death of one of its main proponents and later as people proposed further changes, and the number of current speakers is estimated at between 250 and 5000. However, Ido has proven to be a rich source of Esperanto vocabulary. Some more focused reform projects, affecting only a particular feature of the language, have gained a few adherents. One of these is riism, which modifies the language to incorporate non-sexist language and gender-neutral pronouns. However, most of these projects are specific to individual nationalities (riism from English speakers, for example), and the only changes that have gained acceptance in the Esperanto community have been the minor and gradual bottom-up reforms discussed in the last section. Esperanto is credited with influencing or inspiring several later competing language projects, such as Occidental (1922) and Novial (1928). These always lagged far behind Esperanto in their popularity. By contrast, Interlingua (1951) has greatly surpassed Ido in terms of popularity. It shows little or no Esperanto influence, however. Timeline of Esperanto 1859: L. L. Zamenhof, the creator of Esperanto, is born in Białystok, Russia (now Poland). 1873: The Zamenhof family moves to Warsaw. 1878: Zamenhof celebrates the completion of his universal language project, Lingwe Uniwersala, with high school friends. 1879: Zamenhof attends medical school in Moscow. His father burns his language project while he's away. Meanwhile Schleyer publishes a sketch of Volapük, the first constructed international auxiliary language to acquire a number of speakers. Many Volapük clubs will later switch to Esperanto. 1881: Zamenhof returns to Warsaw to continue medical school, and starts to recreate his project. 1887: Zamenhof marries. In July, with his wife's financial help, he publishes Unua Libro, the first publication introducing Esperanto, in Russian. Polish, German, and French translations are published later that year. 1888: Leo Tolstoy becomes an early supporter. Zamenhof publishes Dua Libro, as well as the first English-language edition of Unua Libro, which proved to be filled with errors. 1889: The second English-language edition of Unua Libro is published in January, translated by Richard H. Geoghegan, and becomes the standard English translation. Henry Phillips, Jr., of the American Philosophical Society, also translates Unua Libro into English. The first volume of La Esperantisto is published in September. The language begins to be called Esperanto. 1894: Zamenhof, reacting to pressure, puts a radical reform to a vote, but it is overwhelmingly rejected. That version of Esperanto is often referred to as Esperanto 1894. 1895: La Esperantisto ceases publication. Lingvo Internacia begins publication in December. 1901: Zamenhof publishes his ideas on a universal religion, based on the philosophy of Hillel the Elder. 1905: Fundamento de Esperanto is published in the spring. The first World Esperanto Congress is held in Boulogne-sur-Mer, with 688 participants and conducted entirely in Esperanto. The Declaration of Boulogne is drafted and ratified at the congress. 1906: The second World Esperanto Congress is held in Geneva, Switzerland, drawing 1200 participants. La Revuo begins publication. 1907: Twelve members of the British parliament nominate Zamenhof for the Nobel Peace Prize. The Ĉekbanko Esperantista (Esperantist Checking Bank) is founded in London, using the spesmilo, an auxiliary Esperanto currency based on the gold standard. A committee organised by Louis Couturat in Paris proposes the Ido reform project, which provides significant competition for Esperanto until the First World War. 1908: The Universal Esperanto Association is founded by Hector Hodler, a 19-year-old Swiss Esperantist. 1909: The International Association of Esperantist Railway Workers is founded in Barcelona. 1910s: Esperanto is taught in state schools in the Republic of China, Samos, and Macedonia. 1910: 42 members of the French parliament nominate Zamenhof for the Nobel Peace Prize. 1914: Lingvo Internacia and La Revuo cease publication. 1917: Zamenhof dies during World War I. 1920: The first Esperanto magazine for the blind, Aŭroro, begins publishing in then-Czechoslovakia. It's still in print today. 1921: The French Academy of the Sciences recommends using Esperanto for international scientific communication. 1922: Esperanto is banned from French schools. 1924: The League of Nations recommends that member states implement Esperanto as an auxiliary language. 1920s: Offices of the Brazilian Ministry of Education use Esperanto for their international correspondence. Lu Xun, the founder of modern Chinese literature, becomes a supporter of Esperanto. Montagu C. Butler is the first to raise Esperanto-speaking children. 1933/34: Reorganisation of the international (neutral) Esperanto movement, under the name UEA. 1934: Encyclopedia of Esperanto first published in Budapest. 1935: Kalocsay and Waringhien publish the influential Plena Gramatiko de Esperanto (Complete Grammar of Esperanto). Esperanto and other planned languages de facto prohibited in Nazi Germany in May. 1936: All Esperanto organisations in Nazi Germany prohibited in June through the prohibition of the UEA and SAT in June by Heinrich Himmler. 1937: Leaders of the Esperanto organisation in the Soviet Union arrested; Esperanto activities made impossible. 1938: The World Esperanto Youth Organisation TEJO is founded. 1939–1945: In World War II many countries are occupied by Germany and the Soviet Union, where Esperanto organisations often were prohibited or Esperanto activities were limited in other ways. 1948: The railway workers' association is refounded as IFEF, the Internacia Fervojista Esperanto-Federacio (International Railway Workers' Esperanto Federation) to foster the use of Esperanto in the administration of the railroads of the world (so far, of Eurasia). 1954: UNESCO establishes consultative relations with the Universal Esperanto Association. 1966: The precursor to Pasporta Servo is launched in Argentina. Pasporta Servo is a global network of Esperanto speakers who host Esperantists traveling through their countries. 1967: István Nemere founds the Renkontiĝo de Esperanto-Familioj, the first organisation for Esperanto-speaking families. 1975: The Esperanto movement spreads to Iran, with three thousand learning the language in Tehran. 1980: The Internacia Junulara Kongreso (International Youth Congress) in Rauma, Finland ratify the Manifesto of Rauma, articulating the view of many in the Esperanto movement that Esperanto is a goal in itself. 1985: UNESCO encourages UN member states to add Esperanto to their school curricula. 1987: 6000 Esperantists attend the 72nd World Esperanto Congress in Warsaw, marking Esperanto's centennial. 1991: The first pan-African Esperanto Conference is held in Lomé, Togo. 1992: PEN International accepts an Esperanto section. 1999: The Esperanto poet William Auld is nominated for the Nobel Prize in Literature. 2001: The Vikipedio project (Esperanto Wikipedia) is launched, resulting in the first general encyclopedia written in a constructed language. It is now one of the most popular websites in Esperanto. 2004: The Europe–Democracy–Esperanto party (E°D°E°) contests the European Parliament elections in France, on a platform of making Esperanto the second language of all EU member states, taking 0.15% of the vote. 2007: Israel issues a stamp to commemorate 120 years of Esperanto (1887–2007). An image of Zamenhof is designed in a text describing his life, reproduced from the Wikipedia article on Esperanto. The corner of the tab shows the flag of the Esperanto movement. 2009: The Senate of Brazil passed a bill which would make Esperanto an optional part of the curriculum in its state schools. As of 2010 the bill has not yet been passed by the Chamber of Deputies. 2015: The 100th World Esperanto Congress is held in Lille, France. Duolingo launches its Esperanto program. 2017: Amikumu is launched, an app connecting Esperantists with other local Esperantists throughout the world. References Bibliography Auld, William. La Fenomeno Esperanto. Rotterdam: UEA, 1988. Dijk, Ziko van. Historio de UEA. Partizánske: Espero, 2012. Forster, Peter Glover. The Esperanto Movement. Den Haag et al. 1982 (Hull 1977). Gobbo, Federico. Is It Possible for All People to Speak the Same Language? The Story of Ludwik Zamenhof and Esperanto (PDF). Korĵenkov, Aleksander. Historio de Esperanto. Kaliningrad 2005. Lins, Ulrich. La Danĝera Lingvo. Gerlingen, Germany: Bleicher Eldonejo, 1988. (Also available in Polish ) Further reading Ronald Creagh & Karine Tourraton. A Short History of Anarchist Participation in the Esperanto Movement. Esperanto Articles containing video clips
10503
https://en.wikipedia.org/wiki/Exile%20%281995%20video%20game%20series%29
Exile (1995 video game series)
Exile is a series of role-playing video games created by Jeff Vogel of Spiderweb Software. They were released as shareware titles for Macintosh and Windows systems. Exile III was also ported to Linux by a third party. There were four games released in the series. All of the games were later revived in the Avernum series. Common to all games in the Exile series are 2D graphics and basic sound. The graphics in the first versions of Exile I and II had simple textures, colours and outlines, which were then replaced in later versions with Exile IIIs graphics. The games are designed to be non-linear and long in gameplay length. Gameplay In each game, the player is required to create six characters to form a party of adventurers or may instead use a default party. The characters' general, combat, magic and miscellaneous skills can be customized along with the character names and graphics. From Exile II onward, characters can have their traits and race configured. The games have three modes including Outdoor Mode, Town Mode and Combat Mode. In Outdoor Mode, the party can enter a town, engage a group of enemies in Combat Mode and rest (if the party has food). In Town Mode, the party can talk to people, purchase from shops (provided they have enough gold), train the characters (provided they have sufficient gold and skill points), find sub-quests, pick up items (from Exile 2 onward, items can be stolen) and enter Combat Mode. In Combat Mode, the party can attack enemies, defend themselves and pick up adjacent items. Combat Mode can be ended at any time in towns, but can only end outdoors when all enemies are killed. Magic is divided into Mage and Priest spell types. Spells for attacking can only be cast in Combat Mode. Some spells, such as Light spells, can be cast at any time. Other spells can only be cast outdoors or in town when not in Combat Mode. Players can create a character equipped with spells up to level 3, but higher level or certain spells must be purchased or found in special encounters. Games Exile I: Escape from the Pit The first game of the Exile trilogy released in January 1995 sees a party of newly created characters thrown from the world above into the subterranean world known as Exile. Once here, the party discovers a civilization formed from the outcasts of the Empire above, a culture beset by constant warfare and monsters galore. The party meets with many of those who wish to get revenge on the Empire for the wrongs it has done to the peoples of Exile. The characters become a rallying point around which the people of Exile who desire vengeance gather to focus their energies into finding a way to strike back against the cruel Emperor of the surface. Together, the party manages to assassinate Emperor Hawthorne in his throne room, banish the demon king Grah-hoth who was becoming a significant threat to the citizens of Exile, and secure an escape route to the surface. Exile II: Crystal Souls The second Exile game follows directly from the first, released in November 1996. The Empire has begun to recognise the threat the Exiles pose and begin sending their army down into Exile in huge numbers. To make matters worse, unknown barriers of energy are sprouting up around the world – sometimes aiding the Exiles, sometimes helping the Empire who can afford the losses much more easily than the Exiles. A new party of characters meets one of the creatures causing the barriers sprouting up in Exile and go to meet with the unknown race to negotiate. In the end, the party is more successful - and the Vahnatai joins with the Exiles to drive out the Empire. With the support of the Vahnatai the Exiles turn the tables on the Empire and successfully repulse their invasion. Exile III: Ruined World The final release in the Exile trilogy takes place some time after Exile II, released in January 1997. The Linux version was ported by Boutell.com in Summer, 2000. A lot of preparation has taken place and now the Exiles are ready to send a selected few back into the light of the surface. However, while the members of the expedition are at first stunned by the sheer beauty of the land around them, they begin to notice that things are not as perfect as they seem. The slimes the party encounters are only the first part of what becomes a series of monsters and terrible occurrences that are blighting the Empire and laying it to waste. While scouting the land, as were the expedition members' orders from the nation of Exile, the members are asked by the Empire to help save the surface from its blight. They bring the Exiles and the Empire together as allies trying to find the cause of the destruction. Blades of Exile Blades of Exile was released in December 1997, consisting of three short scenarios set after the main trilogy as well as an editor that allows players to create their own scenarios, which need not be set in the Exile game world at all. Several hundred custom-made scenarios have been designed since the release of the game in 1997. The most prominent meeting places on the web of the Blades of Exile community are the official company-hosted internet forum. These forums offer support for beginning designers and players, reviews of new scenarios and general discussions about the use of the scenario editor. In June 2007, Jeff Vogel released the source code and game content for Blades of Exile, which is currently under version 2 of the GNU General Public License. Release On December 1, 1998, the first three Exile games also came packaged on a CD called the "Exile Trilogy CD". As of July 8, 2013, these games are freeware on Spiderweb Software's website. Reception Reviews described the Exile trilogy as a "throwback" to old fashioned role-playing games, with deep, complex gameplay and simplistic graphics. Computer Games Strategy Plus rated Exile: Escape from the Pit three and a half out of five stars, calling it "one of the year's best shareware games." The reviewer dropped a half star for its "primitive graphics" but found the game addictive and recommended it to veteran players "hungry for good, old fashioned fantasy role-playing games". Inside Mac Games rated Exile four out of five, calling it an Ultima-style role-playing game that required "patience and thoroughness" and had slightly "cheesy" graphics and sound. Inside Mac Games named Exile as runner up to Heroes of Might and Magic: A Strategic Quest for the best role-playing game of 1996. Inside Mac Games said that Exile II: Crystal Souls had "a size, scope and depth of plot unmatched in Macintosh role-playing games." Exile II: Crystal Souls was an honorable mention for MacUsers award for the best shareware game of 1996, behind winner Escape Velocity. Computer Gaming World said that Exile III: Ruined World appears at first to be "a shareware game with primitive graphics" but reveals itself to be a "remarkably deep" traditional role-playing game with deceptive complexity. The reviewer praised the "well written and witty" NPC dialogue and elegant interface. Emil Pagliarulo reviewed Exile III positively for the Adrenaline Vault, rating the game four out of five stars and calling it an "engrossing, thoroughly entertaining [...] epic computer role-playing game" with simplistic presentation and "enormous depth". Inside Mac Games called Exile III "huge" and "highly detailed", with quality graphics, "an elegant interface", and "one of the largest and most detailed worlds and plots a fantasy role-playing game has offered". Exile III received the 1998 ZDNet Shareware Game of the Year award, selected by the editors of Ziff-Davis magazines FamilyPC, PC Magazine, and Computer Gaming World. Avernum Spiderweb Software has remade the Exile trilogy twice. The games were remade in 2000–2002 as the First Trilogy of the Avernum series, which replaced the two-dimensional tile-based graphics system of Exile with an isometric one and made numerous changes to the RPG system and some changes to the content. The remakes were followed by a remake of Blades of Exile, Blades of Avernum, in 2004. The Exile trilogy was rebooted for a second time with the release of Avernum: Escape from the Pit (2011), Avernum 2: Crystal Souls (2015), and Avernum 3: Ruined World (2018); the remakes featured an enhanced game engine and expanded storylines. Engine and interface While the game engine itself remained relatively similar between all games in the series, the interface went through many changes. Each iteration sported a new layout and color scheme, as well the individual elements, like the inventory and character roster boxes, were also changed to display information differently. Between Exile I and Exile II the most notable difference is a background color change as well as a change to the border of the play window. Between Exile II and Exile III the interface was changed significantly in that the colors and window styles were changed again but the player roster was overhauled and an inventory window was added. The changes between Exile III and Blades of Exile were more subtle and were again of the color and style nature. See also List of open source games References External links Spiderweb Software Blades of Exile Source Code Encyclopedia Ermariana - an extensive encyclopedia of Exile Indie games Linux games Classic Mac OS games Role-playing video games Spiderweb Software games Video game franchises Video games with tile-based graphics Windows games Open-source video games 1997 video games 1996 video games 1995 video games Video game franchises introduced in 1995 Video games developed in the United States
10709
https://en.wikipedia.org/wiki/Geography%20of%20Finland
Geography of Finland
The geography of Finland is characterized by its northern position, its ubiquitous landscapes of intermingled boreal forests and lakes, and its low population density. Finland can be divided into three areas: archipelagoes and coastal lowlands, a slightly higher central lake plateau and uplands to north and northeast. Bordering the Baltic Sea, Gulf of Bothnia, and Gulf of Finland, as well as Sweden, Norway, and Russia, Finland is the northernmost country in the European Union. Most of the population and agricultural resources are concentrated in the south. Northern and eastern Finland are sparsely populated containing vast wilderness areas. Taiga forest is the dominant vegetation type. Size and external boundaries Finland's total area is . Of this area 10% is water, 69% forest, 8% cultivated land and 13% other. Finland is the eighth largest country in Europe after Russia, France, Ukraine, Spain, Sweden, Norway and Germany. As a whole, the shape of Finland's boundaries resembles a figure of a one-armed human. In Finnish, parallels are drawn between the figure and the national personification of Finland – Finnish Maiden (Suomi-neito) – and the country as a whole can be referred in the Finnish language by her name. Even in official context the area around Enontekiö in northwestern part of the country between Sweden and Norway can be referred to as the "Arm" (käsivarsi). After the Continuation War Finland lost major land areas to Russia in the Moscow Armistice of 1944, and the figure was said to have lost the other of her arms, as well as a hem of her "skirt". Relief and geology Geology The bedrock of Finland belong to the Baltic Shield and was formed by a succession of orogenies in Precambrian time. The oldest rocks of Finland, those of Archean age, are found in the east and north. These rocks are chiefly granitoids and migmatitic gneiss. Rocks in central and western Finland originated or came to place during the Svecokarelian orogeny. Following this last orogeny Rapakivi granites intruded various locations of Finland during the Mesoproterozoic and Neoproterozoic, specially at Åland and the southeast. So-called Jotnian sediments occur usually together with Rapakivi granites. The youngest rocks in Finland are those found in the northwestern arm which belong to Scandinavian Caledonides that assembled in Paleozoic times. During the Caledonian orogeny Finland was likely a sunken foreland basin covered by sediments, subsequent uplift and erosion would have eroded all of these sediments. Relief and hydrography About one third of Finland lies below 100 m, and about two thirds lies under 200 m. Finland can be divided into three topographical areas; the coastal landscapes, the interior lake plateau also known as Finnish lake district and Upland Finland. The coastal landscapes are made up mostly of plains below 20 m. These plains tilt gently towards the sea so that where its irregularities surpasses sea-level groups of islands like the Kvarken Archipelago or the Åland Islands are found. Åland is connected to the Finnish mainland by a shallow submarine plateau that does not exceed 20 m in depth. Next to the Gulf of Bothnia the landscape of Finland is extremely flat with height differences no larger than 50 m. This region called the Ostrobothnian Plain extends inland about 100 km and constitute the largest plain in the Nordic countries. The interior lake plateau is dominated by undulating hilly terrain with valley to top height differences of 100 or less and occasionally up to 200 m. Only the area around the lakes Pielinen and Päijänne stand with a subtly more pronounced relief. The relief of the interior lake plateau bears some resemblance to the Swedish Norrland terrain. Upland Finland and areas higher than 200 m are found mostly in the north and east of the country. A limited number of hills and mountains exceed 500 m in height in these regions. Inselberg plains are common in the northern half of the country. In the far north hills reach 200 to 400 m and the landscape is a förfjäll (fore-fell). Only the extreme northwest contains a more dramatic mountain landscape. The subdued landscape of Finland is the result of protracted erosion that has leveled down ancient mountain massifs into near-flat landforms called peneplains. The last major leveling event resulted in the formation of the Sub-Cambrian peneplain in Late Neoproterozoic time. While Finland has remained very close to sea-level since the formation of this last peneplain some further relief was formed by a slight uplift resulting in the carving of valleys by rivers. The slight uplift also means that at parts the uplifted peneplain can be traced as summit accordances. The Quaternary ice ages resulted in the erosion of weak rock and loose materials by glaciers. When the ice masses retreated eroded depressions turned into lakes. Fractures in Finland's bedrock were particularly affected by weathering and erosion, leaving as result trace straight sea and lake inlets. Except a few rivers along the coasts most rivers in Finland drain at some stage into one or more lakes. The drainage basins drain into various directions. Much of Finland drains into the Gulf of Bothnia including the country's largest and longest rivers, Kokemäenjoki and Kemijoki respectively. Finland's largest lake drains by Vuoksi River into Lake Ladoga in Russia. Upland Finland in the east drains east across Russian Republic of Karelia into the White Sea. In the northeast Lake Inari discharges by Paatsjoki into Barents Sea in the Arctic. Quaternary glaciation The ice sheet that covered Finland intermittently during the Quaternary grew out from the Scandinavian Mountains. During the last deglaciation the first parts of Finland to become ice-free, the southeastern coast, did so slightly prior to the Younger Dryas cold-spell 12,700 years before present (BP). The retreat of the ice cover occurred simultaneously from the north-east, the east and southeast. The retreat was fastest from the southeast resulting in the lower course of Tornio being the last part of Finland to be deglaciated. Finally by 10,100 years BP the ice cover had all but left Finland to concentrate in Sweden and Norway before fading away. As the ice sheet became thinner and retreated the land begun to rise by effect of isostacy. Much of Finland was under water when the ice retreated and was gradually uplifted in a process that continues today. Albeit not all areas were drowned at the same time it is estimated at time or another about 62% has been under water. Depending on location in Finland the ancient shoreline reached different maximum heights. In southern Finland 150 to 160 m, in central Finland about 200 m and in eastern Finland up to 220 m. Climate Latitude is the principal influence on Finland's climate. Because of Finland's northern location, winter is the longest season. Only in the south coast is summer as long as winter. On the average, winter lasts from early December to mid March in the archipelago and the southwestern coast and from early October to early May in Lapland. This means that southern portions of the country are snow-covered about three months of the year and the northern, about seven months. The long winter causes about half of the annual of precipitation in the north to fall as snow. Precipitation in the south amounts to about annually. Like that of the north, it occurs all through the year, though not so much of it is snow. The Atlantic Ocean to the west and the Eurasian continent to the east interact to modify the climate of the country. The warm waters of the Gulf Stream and the North Atlantic Drift Current, which warm Norway and Sweden, also warm Finland. Westerly winds bring the warm air currents into the Baltic areas and to the country's shores, moderating winter temperatures, especially in the south. These winds, because of clouds associated with weather systems accompanying the westerlies, also decrease the amount of sunshine received during the summer. By contrast, the continental high pressure system situated over the Eurasian continent counteracts the maritime influences, occasionally causing severe winters and high temperatures in the summer. The highest ever recorded temperature is (Liperi, 29 July 2010). The lowest, (Kittilä, 28 January 1999). The annual middle temperature is relatively high in the southwestern part of the country (), with quite mild winters and warm summers, and low in the northeastern part of Lapland (). Temperature extremes for every month: Extreme highs: January: (January 6, 1973, Mariehamn Airport, Jomala, Åland) February: (February 28, 1943, Ilmala, Helsinki, Uusimaa) March: (March 27, 2007, Helsinki Airport, Vantaa, Uusimaa) April: (April 27, 1921, Jyväskylä, Central Finland) May: (May 30/31, 1995, Ingermaninkylä, Lapinjärvi, Uusimaa) June: (June 24, 1934, Ähtäri, South Ostrobothnia) July: (July 29, 2010, Joensuu Airport, Liperi, North Karelia) August: (August 7, 2010, Heinola, Päijänne Tavastia, and Puumala, South Savo and August 8, 2010, Laune, Lahti, Päijänne Tavastia) September: (September 6, 1968, Rauma, Satakunta) October: (October 14, 2018, Oulu Airport, Oulu, North Ostrobothnia) November: (November 6, 2020, Mariehamn Airport, Jomala, Åland) December: (December 20, 2015, Kokemäki, Satakunta and Pori, Satakunta) Extreme lows: January: (January 28, 1999, Pokka, Kittilä, Lapland) February: (February 5, 1912, Sodankylä, Lapland) March: (March 1, 1971, Tuntsa, Salla, Lapland) April: (April 2, 1912, Kuusamo, North Ostrobothnia and April 9, 1912, Sodankylä, Lapland) May: (May 1, 1971, Kalmankaltio, Enontekiö, Lapland) June: (June 3, 1962, Laanila, Inari, Lapland) July: (July 12, 1958, Kilpisjärvi, Enontekiö, Lapland) August: (August 26, 1980, Naruska, Salla, Lapland) September: (September 26, 1968, Vuotso, Sodankylä, Lapland) October: (October 25, 1968, Sodankylä, Lapland) November: (November 30, 1915, Sodankylä, Lapland) December: (December 21, 1919, Pielisjärvi, North Karelia) Area and boundaries Area: total: land: water: Area – comparative: slightly smaller than Germany, Montana, and Newfoundland and Labrador Land boundaries: total: border countries: Norway , Sweden , Russia Coastline: Maritime claims: Territorial sea: , in the Gulf of Finland; there is a stretch of international waters between Finnish and Estonian claims; Bogskär has separate internal waters and 3 nmi of territorial waters Contiguous zone: Exclusive economic zone: ; extends to continental shelf boundary with Sweden, Estonia, and Russia Continental shelf: depth or to the depth of exploitation Elevation extremes: lowest point: Baltic Sea 0 m highest point: Haltitunturi Resources and land use Natural resources: timber, iron ore, copper, lead, zinc, chromite, nickel, gold, silver, limestone Land use: agricultural land: 7.5% (2018 est.) arable land: 7.4% (2018 est.) permanent crops: 0% (2018 est.) permanent pasture: 0.1% (2018 est.) forest: 72.9% (2018 est.) other: 19.6% (2018 est.) Irrigated land: 690 km2 (2012) Total renewable water resources: 110 billion m3 (2017 est.) Freshwater withdrawal (domestic/industrial/agricultural): municipal: 400 million m3 (2017 est.) industrial: 1.417 billion m3 (2017 est.) agricultural: 50 million m3 (2017 est.) Environmental concerns Natural hazards: Cold periods in winter pose a threat to the unprepared. Environment – current issues: Air pollution from manufacturing and power plants contributing to acid rain; water pollution from industrial wastes, agricultural chemicals; habitat loss threatens wildlife populations. Environment – international agreements: party to: Air Pollution, Air Pollution-Heavy Metals, Air Pollution-Multi-effect Protocol, Air Pollution-Nitrogen Oxides, Air Pollution-Persistent Organic Pollutants (signed 2001, ratified 2002), Air Pollution-Sulphur 85, Air Pollution-Sulphur 94, Air Pollution-Volatile Organic Compounds, Antarctic-Environmental Protection, Antarctic-Marine Living Resources, Antarctic Treaty, Biodiversity, Climate Change, Climate Change-Kyoto Protocol (signed May 1998, ratified together with 14 other EU countries May 31, 2002), Climate Change-Paris Agreement, Comprehensive Nuclear Test Ban, Desertification, Endangered Species, Environmental Modification, Hazardous Wastes, Law of the Sea, Marine Dumping-London Convention, Marine Dumping-London Protocol, Marine Life Conservation, Nuclear Test Ban, Ozone Layer Protection, Ship Pollution, Tropical Timber 2006, Wetlands, Whaling. Other miscellaneous information In Finland there are approximately 168,000 lakes of over in size, and 57,000 of over . A research project by National Land Survey of Finland is currently (2019) seeking to clarify the definition of 'lake' and the number of lakes in Finland. The Finnish capital, Helsinki, is the northernmost capital city on the mainland of any continent, and ranks as second globally (the Icelandic capital Reykjavik takes the first place globally). At , Finland has the second-longest border with Russia of any European country, surpassed only by Ukraine (). The third largest lake, Lake Inari in the Lapland province of extreme northern Finland, has a surface area of , a total shore length of , a maximum depth of , some 3,318 islands, and a total water volume of . Despite its size and numerous recreational opportunities, the lake is scarcely visited sheerly because of its distance from Helsinki, and its daunting distance to other similarly populated areas in the south of the country. Notes References External links
10861
https://en.wikipedia.org/wiki/The%20Trial
The Trial
The Trial (, previously , and ) is a novel written by Franz Kafka in 1914 and 1915 and published posthumously on 26 April 1925. One of his best known works, it tells the story of Josef K., a man arrested and prosecuted by a remote, inaccessible authority, with the nature of his crime revealed neither to him nor to the reader. Heavily influenced by Dostoevsky's Crime and Punishment and The Brothers Karamazov, Kafka even went so far as to call Dostoevsky a blood relative. Like Kafka's two other novels, The Castle and Amerika, The Trial was never completed, although it does include a chapter which appears to bring the story to an intentionally abrupt ending. After Kafka's death in 1924 his friend and literary executor Max Brod edited the text for publication by Verlag Die Schmiede. The original manuscript is held at the Museum of Modern Literature, Marbach am Neckar, Germany. The first English-language translation, by Willa and Edwin Muir, was published in 1937. In 1999, the book was listed in Le Monde 100 Books of the Century and as No. 2 of the Best German Novels of the Twentieth Century. Development Kafka drafted the opening sentence of The Trial in August 1914, and continued work on the novel throughout 1915. This was an unusually productive period for Kafka, despite the outbreak of World War I, which significantly increased the pressures of his day job as an insurance agent. Having begun by writing the opening and concluding sections of the novel, Kafka worked on the intervening scenes in a haphazard manner, using several different notebooks simultaneously. His friend Max Brod, knowing Kafka's habit of destroying his own work, eventually took the manuscript into safekeeping. This manuscript consisted of 161 loose pages torn from notebooks, which Kafka had bundled together into chapters. The order of the chapters was not made clear to Brod, nor was he told which parts were complete and which unfinished. Following Kafka's death in 1924, Brod edited the work and assembled it into a novel to the best of his ability. Further editorial work has been done by later scholars, but Kafka's final vision for The Trial remains unknown. Plot summary On the morning of his thirtieth birthday, Josef K., the chief cashier of a bank, is unexpectedly arrested by two unidentified agents from an unspecified agency for an unspecified crime. Josef is not imprisoned, however, but left "free" and told to await instructions from the Committee of Affairs. Josef's landlady, Frau Grubach, tries to console Josef about the trial, but insinuates that the procedure may be related to an immoral relationship with his neighbor Fräulein Bürstner. Josef visits Bürstner to vent his worries, and then kisses her. A few days later, Josef finds that Fräulein Montag, a lodger from another room, has moved in with Fräulein Bürstner. He suspects that this maneuver is meant to distance him from Bürstner. Josef is ordered to appear at the court's address the coming Sunday, without being told the exact time or room. After a period of exploration, Josef finds the court in the attic. Josef is severely reproached for his tardiness, and he arouses the assembly's hostility after a passionate plea about the absurdity of the trial and the emptiness of the accusation. Josef later tries to confront the presiding judge over his case, but only finds an attendant's wife. The woman gives him information about the process and attempts to seduce him before a law student bursts into the room and takes the woman away, claiming her to be his mistress. The woman's husband then takes Josef on a tour of the court offices, which ends after Josef becomes extremely weak in the presence of other court officials and accused. One evening, in a storage room at his own bank, Josef discovers the two agents who arrested him being whipped by a flogger for asking Josef for bribes and as a result of complaints Josef made at court. Josef tries to argue with the flogger, saying that the men need not be whipped, but the flogger cannot be swayed. The next day he returns to the storage room and is shocked to find everything as he had found it the day before, including the whipper and the two agents. Josef is visited by his uncle, a traveling countryman. Worried by the rumors about his nephew, the uncle introduces Josef to Herr Huld, a sickly and bedridden lawyer tended to by Leni, a young nurse who shows an immediate attraction to Josef. During the conversation, Leni calls Josef away and takes him to the next room for a sexual encounter. Afterward, Josef meets his angry uncle outside, who claims that Josef's lack of respect for the process has hurt his case. During subsequent visits to Huld, Josef realizes that he is a capricious character who will not be of much help. At the bank, one of Josef's clients recommends him to seek the advice of Titorelli, the court's official painter. Titorelli has no real influence within the court, but his deep experience of the process is painfully illuminating to Josef, and he can only suggest complex and unpleasant hypothetical options, as no definitive acquittal has ever been managed. Josef finally decides to dismiss Huld and take control of matters himself. Upon arriving at Huld's office, Josef meets a downtrodden individual, Rudi Block, a client who offers Josef some insight from a client's perspective. Block's case has continued for five years and he has gone from being a successful businessman to being almost bankrupt and is virtually enslaved by his dependence on the lawyer and Leni, with whom he appears to be sexually involved. The lawyer mocks Block in front of Josef for his dog-like subservience. This experience further poisons Josef's opinion of his lawyer. Josef is put in charge of accompanying an important Italian client to the city's cathedral. While inside the cathedral, a priest calls Josef by name and tells him a fable (which was published earlier as "Before the Law") that is meant to explain his situation. The priest tells Josef that the parable is an ancient text of the court, and many generations of court officials have interpreted it differently. On the eve of Josef's thirty-first birthday, two men arrive at his apartment to execute him. They lead him to a small quarry outside the city, and kill him with a butcher's knife. Josef summarizes his situation with his last words: "Like a dog!" Characters Josef K. – The tale's protagonist: a thirty year old, unmarried bank administrator living in an unnamed city. Fräulein Bürstner – A boarder in the same house as Josef K. She lets him kiss her one night, but then rebuffs his advances. K. briefly catches sight of her, or someone who looks similar to her, in the final pages of the novel. Fräulein Montag – Friend of Fräulein Bürstner, she talks to K. about ending his relationship with Fräulein Bürstner after his arrest. She claims she can bring him insight, because she is an objective third party. Willem and Franz – Officers who arrest K. one morning but refuse to disclose the crime he is said to have committed. They are later flogged. Inspector – Man who conducts a proceeding at Josef K.'s boardinghouse to inform K. officially that he is under arrest. Rabinsteiner, Kullich and Kaminer – Junior bank employees who attend the proceeding at the boardinghouse. Frau Grubach – The proprietress of the lodging house in which K. lives. She holds K. in high esteem, despite his arrest. Woman in the Court – The first judgement of K happens in her house. She claims help from K. because she does not want to be abused by the magistrates. Student – Deformed man who acts under orders of the instruction judge. Will be a powerful man in the future. Instruction Judge – First Judge of K. In his trial, he confuses K. with a wall painter. Uncle Karl – K.'s impetuous uncle from the country, formerly his guardian. Upon hearing of the trial, Karl insists that K. hire Herr Huld, the lawyer. Herr Huld, the Lawyer – K.'s pompous and pretentious advocate who provides little in the way of action and far too much in the way of anecdote. Leni – Herr Huld's nurse, she has feelings for Josef K. and soon becomes his lover. She shows him her webbed hand, yet another reference to the motif of the hand throughout the book. Apparently, she finds accused men extremely attractive — the fact of their indictment makes them irresistible to her. Albert – Office director at the court, and a friend of Huld. Flogger – Man who punishes Franz and Willem in the Bank after K.'s complaints against the two agents in his first Judgement. Vice-President – K.'s unctuous rival at the Bank, only too willing to catch K. in a compromising situation. He repeatedly takes advantage of K.'s preoccupation with the trial to advance his own ambitions. President – Manager of the Bank. A sickly figure, whose position the Vice-President is trying to assume. Gets on well with K., inviting him to various engagements. Rudi Block, the Merchant – Block is another accused man and client of Huld. His case is five years old, and he is but a shadow of the prosperous grain dealer he once was. All his time, energy, and resources are now devoted to his case, to the detriment of his own life. Although he has hired five additional lawyers on the side, he is completely and pathetically subservient to Huld. Manufacturer – Person who hears about K.'s case and advises him to see a painter who knows how the court system works. Titorelli, the Painter – Titorelli inherited the position of Court Painter from his father. He knows a great deal about the comings and goings of the court's lowest level. He offers to help K., and manages to unload a few identical landscape paintings on the accused man. Priest – Prison chaplain whom K. encounters in a cathedral. The priest advises K. that his case is going badly and tells him to accept his fate. Doorkeeper and Farmer – The characters of the Chaplain's Tale. Translations into English Everyman's Library, 30 June 1992, Translation: Willa and Edwin Muir, Schocken Books, 25 May 1999, Translation: Breon Mitchell, Translator's preface is available online Dover Thrift Editions, 22 July 2009, Translation: David Wyllie, Oxford World's Classics, 4 October 2009, Translation: Mike Mitchell, Penguin Modern Classics, 29 June 2000, Translation: Idris Parry, , 15 September 2012, Translation: Susanne Lück and Maureen Fitzgibbons, BigFontBooks.com, Large Print Edition, 4 June 2019, Translation: David Wyllie, In addition, a graphic novel adaptation by Chantal Montellier (illustrations) and David Zane Mairowitz (adaptation) appeared on April 15, 2008. Audiobooks in English Blackstone Audio, Narrated by Geoffrey Howard, Translated by Breon Mitchell, 7 hours 47 minutes, Published August 4, 2008 Dramatic adaptations Stage The writer and director Steven Berkoff adapted several of Kafka's novels into plays and directed them for stage. His version of The Trial was first performed in 1970 in London and published in 1981. Israeli director Rina Yerushalmi adapted The Trial (paired with Samuel Beckett's Malone Dies) for a production called Ta, Ta, Tatata presented in June 1970 at La MaMa Experimental Theatre Club. Chicago based writer, Greg Allen, wrote and directed K., based on The Trial. After award-winning runs in Chicago and New York, it was produced by The Hypocrites and ran for several months in 2010 at The Chopin Theater in Chicago. Joseph K, written by Tom Basden and based on The Trial, takes place in modern-day London, with the protagonist cast as a City banker. It ran at the Gate Theatre, Notting Hill, London, in late 2010. Gottfried von Einem wrote an opera, Der Prozeß, based on the novel. Its American debut was directed by Otto Preminger. The writer Serge Lamothe adapted The Trial for the stage. Directed by François Girard, his version of The Trial was first performed in 2004 in Montreal and Ottawa, Canada, and published in 2005. Between June and August 2015 The Young Vic theatre in London staged a version of The Trial adapted by Nick Gill and starring Rory Kinnear as K. Jean-Louis Barrault and Andre Gide adapted the novel for the stage, performed in Paris in 1947. An operatic adaptation of The Trial by the composer Philip Glass was premiered by Music Theatre Wales in October 2014. Radio On 19 May 1946, Columbia Workshop broadcast an adaptation of The Trial by Davidson Taylor with an original musical score by Bernard Herrmann and starring Karl Swenson as Joseph K. In 1982, Mike Gwilym starred as Josef K. with Miriam Margolyes as Leni in an adaptation on BBC Radio 4 dramatised for radio by Hanif Kureishi. Sam Troughton starred as Joseph Kay in a new adaptation by Mark Ravenhill titled The Process directed by Polly Thomas and broadcast on 10 May 2015 on BBC Radio 3's Drama on 3 program. Film In the 1962 film adaptation by Orson Welles, Josef K. is played by Anthony Perkins and The Advocate by Welles. Martin Scorsese's 1985 film After Hours contains a scene adapted from "Before the Law", in which the protagonist is trying to get into a nightclub named Club Berlin. The 1993 film The Trial was based on Harold Pinter's screenplay adaptation. Directed by David Jones, it starred Kyle MacLachlan as Josef K. and Anthony Hopkins as The Priest. The 2017 science fiction film Blade Runner 2049 features a protagonist named Officer K, who is referred to as "Joe", an allusion to Josef K. References Further reading External links Der Prozeß, original text in German Kafka's parable "Before the Law", Herbert Deinert, May 1964, Cornell University 1925 German-language novels 1925 science fiction novels Absurdist fiction 1925 Austrian novels Novels published posthumously Austrian novels adapted into films Dystopian novels Existentialist novels Fictional lawsuits Modernist novels Novels adapted into operas Novels by Franz Kafka Unfinished novels Bureaucracy in fiction
11012
https://en.wikipedia.org/wiki/Forth%20%28programming%20language%29
Forth (programming language)
Forth is a procedural, stack-oriented programming language and interactive environment designed by Charles H. "Chuck" Moore and first used by other programmers in 1970. Although not an acronym, the language's name in its early years was often spelled in all capital letters as FORTH. The FORTH-79 and FORTH-83 implementations, which were not written by Moore, became de facto standards, and an official standardization of the language was published in 1994 as ANS Forth. A wide range of Forth derivatives existed before and after ANS Forth. The free software Gforth implementation is actively maintained, as are several commercially supported systems. Forth typically combines a compiler with an integrated command shell, where the user interacts via subroutines called words. Words can be defined, tested, redefined, and debugged without recompiling or restarting the whole program. All syntactic elements, including variables, operators, and control flow, are defined as words. A stack is used to pass parameters between words, leading to a Reverse Polish Notation style. For much of Forth's existence, the standard technique was to compile to threaded code, which can be interpreted faster than bytecode. One of the early benefits of Forth was size: an entire development environment—including compiler, editor, and user programs—could fit in memory on an 8-bit or similarly limited system. No longer constrained by space, there are modern implementations that generate optimized machine code like other language compilers. The relative simplicity of creating a basic Forth system has led to many personal and proprietary variants, such as the custom Forth used to implement the bestselling 1986 video game Starflight from Electronic Arts. Forth is used in the Open Firmware boot loader, in space applications such as the Philae spacecraft, and in other embedded systems which involve interaction with hardware. Moore later developed a series of microprocessors for executing compiled Forth-like code directly and experimented with smaller languages based on Forth concepts, including cmForth and colorForth. Most of these languages were designed to support Moore's own projects, such as chip design. Uses Forth has a niche in astronomical and space applications as well as a history in embedded systems. The Open Firmware boot ROMs used by Apple, IBM, Sun, and OLPC XO-1 contain a Forth environment. Forth has often been used to bring up new hardware. Forth was the first resident software on the new Intel 8086 chip in 1978, and MacFORTH was the first resident development system for the Macintosh 128K in 1984. Atari, Inc. used an elaborate animated demo written in Forth to showcase capabilities of the Atari 400 and 800 computers in department stores. Three video games from Electronic Arts, published in the 1980s, were written in Forth: Worms? (1983), Starflight (1986), and Lords of Conquest (1986). Robot coding game ChipWits (1984) was written in MacFORTH. Ashton-Tate's RapidFile (1986), a flat-file database program, and VP-Planner from Paperback Software International (1983), a spreadsheet program competing with Lotus 1-2-3, were written in Forth. The Canon Cat (1987) uses Forth for its system programming. Rockwell produced single-chip microcomputers with resident Forth kernels: the R65F11 and R65F12. ASYST was a Forth expansion for measuring and controlling on PCs. History Forth evolved from Charles H. Moore's personal programming system, which had been in continuous development since 1968. Forth was first exposed to other programmers in the early 1970s, starting with Elizabeth Rather at the United States National Radio Astronomy Observatory (NRAO). After their work at NRAO, Charles Moore and Elizabeth Rather formed FORTH, Inc. in 1973, refining and porting Forth systems to dozens of other platforms in the next decade. Forth is so-named, because in 1968 "the file holding the interpreter was labeled FOURTH, for 4th (next) generation software, but the IBM 1130 operating system restricted file names to five characters." Moore saw Forth as a successor to compile-link-go third-generation programming languages, or software for "fourth generation" hardware. FORTH, Inc.'s microFORTH was developed for the Intel 8080, Motorola 6800, Zilog Z80, and RCA 1802 microprocessors, starting in 1976. MicroFORTH was later used by hobbyists to generate Forth systems for other architectures, such as the 6502 in 1978. The Forth Interest Group was formed in 1978. It promoted and distributed its own version of the language, FIG-Forth, for most makes of home computer. Forth was popular in the early 1980s, because it was well suited to the limited memory of microcomputers. The ease of implementing the language led to many implementations. The British Jupiter ACE home computer has Forth in its ROM-resident operating system. Insoft GraFORTH is a version of Forth with graphics extensions for the Apple II. Common practice was codified in the de facto standards FORTH-79 and FORTH-83 in the years 1979 and 1983, respectively. These standards were unified by ANSI in 1994, commonly referred to as Forth. As of 2018, the source for the original 1130 version of FORTH has been recovered, and is now being updated to run on a restored or emulated 1130 system. Overview Forth emphasizes the use of small, simple functions called words. Words for bigger tasks call upon many smaller words that each accomplish a distinct sub-task. A large Forth program is a hierarchy of words. These words, being distinct modules that communicate implicitly via a stack mechanism, can be prototyped, built and tested independently. The highest level of Forth code may resemble an English-language description of the application. Forth has been called a meta-application language: a language that can be used to create problem-oriented languages. Forth relies on explicit use of a data stack and reverse Polish notation which is commonly used in calculators from Hewlett-Packard. In RPN, the operator is placed after its operands, as opposed to the more common infix notation where the operator is placed between its operands. Postfix notation makes the language easier to parse and extend; Forth's flexibility makes a static BNF grammar inappropriate, and it does not have a monolithic compiler. Extending the compiler only requires writing a new word, instead of modifying a grammar and changing the underlying implementation. Using RPN, one can get the result of the mathematical expression (25 * 10 + 50) this way: 25 10 * 50 + CR . 300 ok First the numbers 25 and 10 are put on the stack. The word * takes the top two numbers from the stack, multiplies them, and puts the product back on the stack. Then the number 50 is placed on the stack. The word + adds the top two values, pushing the sum. CR (carriage return) starts the output on a new line. Finally, . prints the result. As everything has completed successfully, the Forth system prints OK. Even Forth's structural features are stack-based. For example: : FLOOR5 ( n -- n' ) DUP 6 < IF DROP 5 ELSE 1 - THEN ; The colon indicates the beginning of a new definition, in this case a new word (again, word is the term used for a subroutine) called FLOOR5. The text in parentheses is a comment, advising that this word expects a number on the stack and will return a possibly changed number (on the stack). The subroutine uses the following commands: DUP duplicates the number on the stack; 6 pushes a 6 on top of the stack; < compares the top two numbers on the stack (6 and the DUPed input), and replaces them with a true-or-false value; IF takes a true-or-false value and chooses to execute commands immediately after it or to skip to the ELSE; DROP discards the value on the stack; 5 pushes a 5 on top of the stack; and THEN ends the conditional. The FLOOR5 word is equivalent to this function written in the C programming language using the ternary operator '?:' int floor5(int v) { return (v < 6) ? 5 : (v - 1); } This function is written more succinctly as: : FLOOR5 ( n -- n' ) 1- 5 MAX ; This can be run as follows: 1 FLOOR5 CR . 5 ok 8 FLOOR5 CR . 7 ok First a number (1 or 8) is pushed onto the stack, FLOOR5 is called, which pops the number again and pushes the result. CR moves the output to a new line (again, this is only here for readability). Finally, a call to . pops the result and prints. Facilities Forth's grammar has no official specification. Instead, it is defined by a simple algorithm. The interpreter reads a line of input from the user input device, which is then parsed for a word using spaces as a delimiter; some systems recognise additional whitespace characters. When the interpreter finds a word, it looks the word up in the dictionary. If the word is found, the interpreter executes the code associated with the word, and then returns to parse the rest of the input stream. If the word isn't found, the word is assumed to be a number and an attempt is made to convert it into a number and push it on the stack; if successful, the interpreter continues parsing the input stream. Otherwise, if both the lookup and the number conversion fail, the interpreter prints the word followed by an error message indicating that the word is not recognised, flushes the input stream, and waits for new user input. The definition of a new word is started with the word : (colon) and ends with the word ; (semi-colon). For example, : X DUP 1+ . . ; will compile the word X, and makes the name findable in the dictionary. When executed by typing 10 X at the console this will print 11 10. Most Forth systems include an assembler to write words using the processor's facilities. Forth assemblers often use a reverse Polish syntax in which the parameters of an instruction precede the instruction. A typical reverse Polish assembler prepares the operands on the stack and the mnemonic copies the whole instruction into memory as the last step. A Forth assembler is by nature a macro assembler, so that it is easy to define an alias for registers according to their role in the Forth system: e.g. "dsp" for the register used as the data stack pointer. Operating system, files, and multitasking Most Forth systems run under a host operating system such as Microsoft Windows, Linux or a version of Unix and use the host operating system's file system for source and data files; the ANSI Forth Standard describes the words used for I/O. All modern Forth systems use normal text files for source, even if they are embedded. An embedded system with a resident compiler gets its source via a serial line. Classic Forth systems traditionally use neither operating system nor file system. Instead of storing code in files, source code is stored in disk blocks written to physical disk addresses. The word BLOCK is employed to translate the number of a 1K-sized block of disk space into the address of a buffer containing the data, which is managed automatically by the Forth system. Block use has become rare since the mid-1990s. In a hosted system those blocks too are allocated in a normal file in any case. Multitasking, most commonly cooperative round-robin scheduling, is normally available (although multitasking words and support are not covered by the ANSI Forth Standard). The word PAUSE is used to save the current task's execution context, to locate the next task, and restore its execution context. Each task has its own stacks, private copies of some control variables and a scratch area. Swapping tasks is simple and efficient; as a result, Forth multitaskers are available even on very simple microcontrollers, such as the Intel 8051, Atmel AVR, and TI MSP430. Other non-standard facilities include a mechanism for issuing calls to the host OS or windowing systems, and many provide extensions that employ the scheduling provided by the operating system. Typically they have a larger and different set of words from the stand-alone Forth's PAUSE word for task creation, suspension, destruction and modification of priority. Self-compilation and cross compilation A full-featured Forth system with all source code will compile itself, a technique commonly called meta-compilation or self-hosting, by Forth programmers (although the term doesn't exactly match meta-compilation as it is normally defined). The usual method is to redefine the handful of words that place compiled bits into memory. The compiler's words use specially named versions of fetch and store that can be redirected to a buffer area in memory. The buffer area simulates or accesses a memory area beginning at a different address than the code buffer. Such compilers define words to access both the target computer's memory, and the host (compiling) computer's memory. After the fetch and store operations are redefined for the code space, the compiler, assembler, etc. are recompiled using the new definitions of fetch and store. This effectively reuses all the code of the compiler and interpreter. Then, the Forth system's code is compiled, but this version is stored in the buffer. The buffer in memory is written to disk, and ways are provided to load it temporarily into memory for testing. When the new version appears to work, it is written over the previous version. Numerous variations of such compilers exist for different environments. For embedded systems, the code may instead be written to another computer, a technique known as cross compilation, over a serial port or even a single TTL bit, while keeping the word names and other non-executing parts of the dictionary in the original compiling computer. The minimum definitions for such a Forth compiler are the words that fetch and store a byte, and the word that commands a Forth word to be executed. Often the most time-consuming part of writing a remote port is constructing the initial program to implement fetch, store and execute, but many modern microprocessors have integrated debugging features (such as the Motorola CPU32) that eliminate this task. Structure of the language The basic data structure of Forth is the "dictionary" which maps "words" to executable code or named data structures. The dictionary is laid out in memory as a tree of linked lists with the links proceeding from the latest (most recently) defined word to the oldest, until a sentinel value, usually a NULL pointer, is found. A context switch causes a list search to start at a different leaf. A linked list search continues as the branch merges into the main trunk leading eventually back to the sentinel, the root. There can be several dictionaries. In rare cases such as meta-compilation a dictionary might be isolated and stand-alone. The effect resembles that of nesting namespaces and can overload keywords depending on the context. A defined word generally consists of head and body with the head consisting of the name field (NF) and the link field (LF), and body consisting of the code field (CF) and the parameter field (PF). Head and body of a dictionary entry are treated separately because they may not be contiguous. For example, when a Forth program is recompiled for a new platform, the head may remain on the compiling computer, while the body goes to the new platform. In some environments (such as embedded systems) the heads occupy memory unnecessarily. However, some cross-compilers may put heads in the target if the target itself is expected to support an interactive Forth. The exact format of a dictionary entry is not prescribed, and implementations vary. Structure of the compiler The compiler itself is not a monolithic program. It consists of Forth words visible to the system, and usable by a programmer. This allows a programmer to change the compiler's words for special purposes. The "compile time" flag in the name field is set for words with "compile time" behavior. Most simple words execute the same code whether they are typed on a command line, or embedded in code. When compiling these, the compiler simply places code or a threaded pointer to the word. The classic examples of compile-time words are the control structures such as IF and WHILE. Almost all of Forth's control structures and almost all of its compiler are implemented as compile-time words. Apart from some rarely used control flow words only found in a few implementations, such as the conditional return word ?EXIT used in Ulrich Hoffmann's preForth, all of Forth's control flow words are executed during compilation to compile various combinations of primitive words along with their branch addresses. For instance, IF and WHILE, and the words that match with those, set up BRANCH (unconditional branch) and ?BRANCH (pop a value off the stack, and branch if it is false). Counted loop control flow words work similarly but set up combinations of primitive words that work with a counter, and so on. During compilation, the data stack is used to support control structure balancing, nesting, and back-patching of branch addresses. The snippet: ... DUP 6 < IF DROP 5 ELSE 1 - THEN ... would often be compiled to the following sequence inside a definition: ... DUP LIT 6 < ?BRANCH 5 DROP LIT 5 BRANCH 3 LIT 1 - ... The numbers after BRANCH represent relative jump addresses. LIT is the primitive word for pushing a "literal" number onto the data stack. (Faster, shorter code would be compiled using pointers to constants instead of LIT and embedded data, if any of the numbers involved have been separately defined as constants. There would be similar changes if yet other words were used instead of constants, and so on.) Compilation state and interpretation state The word : (colon) parses a name as a parameter, creates a dictionary entry (a colon definition) and enters compilation state. The interpreter continues to read space-delimited words from the user input device. If a word is found, the interpreter executes the compilation semantics associated with the word, instead of the interpretation semantics. The default compilation semantics of a word are to append its interpretation semantics to the current definition. The word ; (semi-colon) finishes the current definition and returns to interpretation state. It is an example of a word whose compilation semantics differ from the default. The interpretation semantics of ; (semi-colon), most control flow words, and several other words are undefined in Forth, meaning that they must only be used inside of definitions and not on the interactive command line. The interpreter state can be changed manually with the words [ (left-bracket) and ] (right-bracket) which enter interpretation state or compilation state, respectively. These words can be used with the word LITERAL to calculate a value during a compilation and to insert the calculated value into the current colon definition. LITERAL has the compilation semantics to take an object from the data stack and to append semantics to the current colon definition to place that object on the data stack. In Forth, the current state of the interpreter can be read from the flag STATE which contains the value true when in compilation state and false otherwise. This allows the implementation of so-called state-smart words with behavior that changes according to the current state of the interpreter. Immediate words The word IMMEDIATE marks the most recent colon definition as an immediate word, effectively replacing its compilation semantics with its interpretation semantics. Immediate words are normally executed during compilation, not compiled, but this can be overridden by the programmer in either state. ; is an example of an immediate word. In Forth, the word POSTPONE takes a name as a parameter and appends the compilation semantics of the named word to the current definition even if the word was marked immediate. Forth-83 defined separate words COMPILE and [COMPILE] to force the compilation of non-immediate and immediate words, respectively. Instead of reserving space for an Immediate flag in every definition, some implementations of Forth use an Immediates Dictionary which is checked first when in compile mode. Unnamed words and execution tokens In Forth, unnamed words can be defined with the word :NONAME which compiles the following words up to the next ; (semi-colon) and leaves an execution token on the data stack. The execution token provides an opaque handle for the compiled semantics, similar to the function pointers of the C programming language. Execution tokens can be stored in variables. The word EXECUTE takes an execution token from the data stack and performs the associated semantics. The word COMPILE, (compile-comma) takes an execution token from the data stack and appends the associated semantics to the current definition. The word ' (tick) takes the name of a word as a parameter and returns the execution token associated with that word on the data stack. In interpretation state, ' RANDOM-WORD EXECUTE is equivalent to RANDOM-WORD. Parsing words and comments The words : (colon), POSTPONE, ' (tick) are examples of parsing words that take their arguments from the user input device instead of the data stack. Another example is the word ( (paren) which reads and ignores the following words up to and including the next right parenthesis and is used to place comments in a colon definition. Similarly, the word \ (backslash) is used for comments that continue to the end of the current line. To be parsed correctly, ( (paren) and \ (backslash) must be separated by whitespace from the following comment text. Structure of code In most Forth systems, the body of a code definition consists of either machine language, or some form of threaded code. The original Forth which follows the informal FIG standard (Forth Interest Group), is a TIL (Threaded Interpretive Language). This is also called indirect-threaded code, but direct-threaded and subroutine threaded Forths have also become popular in modern times. The fastest modern Forths, such as SwiftForth, VFX Forth, and iForth, compile Forth to native machine code. Data objects When a word is a variable or other data object, the CF points to the runtime code associated with the defining word that created it. A defining word has a characteristic "defining behavior" (creating a dictionary entry plus possibly allocating and initializing data space) and also specifies the behavior of an instance of the class of words constructed by this defining word. Examples include: VARIABLE Names an uninitialized, one-cell memory location. Instance behavior of a VARIABLE returns its address on the stack. CONSTANT Names a value (specified as an argument to CONSTANT). Instance behavior returns the value. CREATE Names a location; space may be allocated at this location, or it can be set to contain a string or other initialized value. Instance behavior returns the address of the beginning of this space. Forth also provides a facility by which a programmer can define new application-specific defining words, specifying both a custom defining behavior and instance behavior. Some examples include circular buffers, named bits on an I/O port, and automatically indexed arrays. Data objects defined by these and similar words are global in scope. The function provided by local variables in other languages is provided by the data stack in Forth (although Forth also has real local variables). Forth programming style uses very few named data objects compared with other languages; typically such data objects are used to contain data which is used by a number of words or tasks (in a multitasked implementation). Forth does not enforce consistency of data type usage; it is the programmer's responsibility to use appropriate operators to fetch and store values or perform other operations on data. Examples “Hello, World!” : HELLO ( -- ) CR ." Hello, World!" ; HELLO <cr> Hello, World! The word CR (Carriage Return) causes the following output to be displayed on a new line. The parsing word ." (dot-quote) reads a double-quote delimited string and appends code to the current definition so that the parsed string will be displayed on execution. The space character separating the word ." from the string Hello, World! is not included as part of the string. It is needed so that the parser recognizes ." as a Forth word. A standard Forth system is also an interpreter, and the same output can be obtained by typing the following code fragment into the Forth console: CR .( Hello, World!) .( (dot-paren) is an immediate word that parses a parenthesis-delimited string and displays it. As with the word ." the space character separating .( from Hello, World! is not part of the string. The word CR comes before the text to print. By convention, the Forth interpreter does not start output on a new line. Also by convention, the interpreter waits for input at the end of the previous line, after an ok prompt. There is no implied "flush-buffer" action in Forth's CR, as sometimes is in other programming languages. Mixing states of compiling and interpreting Here is the definition of a word EMIT-Q which when executed emits the single character Q: : EMIT-Q 81 ( the ASCII value for the character 'Q' ) EMIT ; This definition was written to use the ASCII value of the Q character (81) directly. The text between the parentheses is a comment and is ignored by the compiler. The word EMIT takes a value from the data stack and displays the corresponding character. The following redefinition of EMIT-Q uses the words [ (left-bracket), ] (right-bracket), CHAR and LITERAL to temporarily switch to interpreter state, calculate the ASCII value of the Q character, return to compilation state and append the calculated value to the current colon definition: : EMIT-Q [ CHAR Q ] LITERAL EMIT ; The parsing word CHAR takes a space-delimited word as parameter and places the value of its first character on the data stack. The word [CHAR] is an immediate version of CHAR. Using [CHAR], the example definition for EMIT-Q could be rewritten like this: : EMIT-Q [CHAR] Q EMIT ; \ Emit the single character 'Q' This definition used \ (backslash) for the describing comment. Both CHAR and [CHAR] are predefined in Forth. Using IMMEDIATE and POSTPONE, [CHAR] could have been defined like this: : [CHAR] CHAR POSTPONE LITERAL ; IMMEDIATE A complete RC4 cipher program In 1987, Ron Rivest developed the RC4 cipher-system for RSA Data Security, Inc. Its description follows: The following Standard Forth version uses Core and Core Extension words only. 0 value ii 0 value jj 0 value KeyAddr 0 value KeyLen create SArray 256 allot \ state array of 256 bytes : KeyArray KeyLen mod KeyAddr ; : get_byte + c@ ; : set_byte + c! ; : as_byte 255 and ; : reset_ij 0 TO ii 0 TO jj ; : i_update 1 + as_byte TO ii ; : j_update ii SArray get_byte + as_byte TO jj ; : swap_s_ij jj SArray get_byte ii SArray get_byte jj SArray set_byte ii SArray set_byte ; : rc4_init ( KeyAddr KeyLen -- ) 256 min TO KeyLen TO KeyAddr 256 0 DO i i SArray set_byte LOOP reset_ij BEGIN ii KeyArray get_byte jj + j_update swap_s_ij ii 255 < WHILE ii i_update REPEAT reset_ij ; : rc4_byte ii i_update jj j_update swap_s_ij ii SArray get_byte jj SArray get_byte + as_byte SArray get_byte xor ; This is one of many ways to test the code: hex create AKey 61 c, 8A c, 63 c, D2 c, FB c, : test cr 0 DO rc4_byte . LOOP cr ; AKey 5 rc4_init 2C F9 4C EE DC 5 test \ output should be: F1 38 29 C9 DE Implementations Because Forth is simple to implement and has no standard reference implementation, there are numerous versions of the language. In addition to supporting the standard varieties of desktop computer systems (POSIX, Microsoft Windows, macOS), many of these Forth systems also target a variety of embedded systems. Listed here are some of the systems which conform to the 1994 Forth standard. ASYST, a Forth-like system for data collection and analysis Gforth, a portable Forth implementation from the GNU Project noForth, an ANS Forth implementation (as far as possible) for Flash microcontrollers (MSP430 & Risc-V) Open Firmware, a bootloader and Firmware standard based on Forth pForth, portable Forth written in C SP-Forth, Forth implementation from the Russian Forth Interest Group (RuFIG) Swift Forth, machine code generating implementation from Forth, Inc. VFX Forth, optimizing native code Forth See also RTX2010, a CPU that runs Forth natively Notes References Further reading External links Programming a problem-oriented language, an unpublished book, by Charles H. Moore, June 1970 Forth programming language family Concatenative programming languages Stack-based virtual machines Systems programming languages Programming languages created in 1970 Extensible syntax programming languages Programming languages with an ISO standard Programming languages 1970 software Articles with example C code
11223
https://en.wikipedia.org/wiki/Four%20Weddings%20and%20a%20Funeral
Four Weddings and a Funeral
Four Weddings and a Funeral is a 1994 British romantic comedy film directed by Mike Newell. It is the first of several films by screenwriter Richard Curtis to feature Hugh Grant, and follows the adventures of Charles (Grant) and his circle of friends through a number of social occasions as they each encounter romance. Andie MacDowell stars as Charles's love interest Carrie, with Kristin Scott Thomas, James Fleet, Simon Callow, John Hannah, Charlotte Coleman, David Bower, Corin Redgrave, and Rowan Atkinson in supporting roles. The film was made in six weeks, cost under £3 million, and became an unexpected success and the highest-grossing British film in history at the time, with worldwide box office total of $245.7 million, and receiving Academy Award nominations for Best Picture and Best Original Screenplay. Additionally, Grant won the Golden Globe Award for Best Actor - Motion Picture Musical or Comedy and the BAFTA Award for Best Actor in a Leading Role, and the film won the BAFTA Awards Best Film, Best Direction, and Best Actress in a Supporting Role for Scott Thomas. The film's success propelled Hugh Grant to international stardom, particularly in the United States. In 1999, Four Weddings and a Funeral was placed 23rd on the British Film Institute's 100 greatest British films of the 20th century. In 2016, Empire magazine ranked it 21st in their list of the 100 best British films. A 2017 poll of 150 actors, directors, writers, producers, and critics for Time Out magazine ranked it the 74th best British film ever. Curtis reunited director Newell and the surviving cast for a 25th anniversary reunion Comic Relief short entitled One Red Nose Day and a Wedding, which aired in the UK during Red Nose Day on 15 March 2019. Plot At the wedding of Angus and Laura in Somerset, the unmarried best man Charles; his flatmate Scarlett; his friend Fiona and her brother, Tom; Gareth and his partner Matthew; and Charles's deaf brother, David, endure the festivities. At the reception, Charles is attracted to Carrie, a young American who has been working in England. They spend the night together. In the morning, Carrie, who is returning to the U.S., laments they may have "missed a great opportunity". Three months later at the London wedding of Bernard and Lydia – who became an item at the previous wedding – Charles is excited to run into Carrie, who has returned to the U.K. He is quickly disappointed after meeting the snobbish Hamish, Carrie's much older Scottish fiancé. During the reception, Charles experiences further humiliation from several ex-girlfriends, including the distraught Henrietta, who angrily claims Charles is a "serial monogamist" and fears commitment. He retreats to an empty hotel suite where he sees Carrie and Hamish depart by taxi. He becomes trapped in the room when the newlyweds stumble in to have sex. Carrie returns to the reception, and she and Charles spend another night together. A month later, Charles receives an invitation to Carrie's wedding. While shopping for a gift, he runs into her. While Charles helps Carrie look for a wedding dress, she recounts her 33 sexual partners; Charles, who was number 32, awkwardly confesses he loves her, which Carrie reluctantly rebuffs. A month later, Charles and his friends attend Carrie and Hamish's wedding in Perthshire. The gregarious Gareth instructs the group to seek potential mates; Scarlett meets Chester, an American from Texas. As Charles watches Carrie and Hamish dance, Fiona deduces his heartbreak. She tells him that she remains single because she has long been in love with him; though sympathetic, Charles does not reciprocate her feelings. During Hamish's toast, Gareth suffers a fatal heart attack. At Gareth's funeral, Matthew recites "Funeral Blues", a poem by British-American poet W. H. Auden. Carrie and Charles share a brief moment and, later, Charles and Tom ponder that despite their clique's pride in being single, Gareth and Matthew were as a "married" couple. They wonder whether the search for "one true love" is futile. Ten months later, Charles's own wedding day arrives; his bride is Henrietta. While seating guests, Tom meets and is immediately smitten with his distant cousin, Deirdre, whom he has not seen since childhood. Carrie arrives and tells Charles that she and Hamish have separated following a difficult marriage. Charles, stunned, privately has an emotional crisis. After brother David and Matthew counsel him, Charles resolves to proceed with the wedding. During the ceremony, the vicar asks if anyone present has any reason why the couple should not marry; David interrupts and, using sign language, says the groom has doubts and loves someone else. Charles confirms this, and Henrietta angrily punches him, halting the ceremony. A few hours later, Carrie arrives at Charles' flat and apologizes for causing the fiasco. Charles says he loves Carrie and proposes a lifelong commitment without marriage, which Carrie accepts. As they kiss, a thunderbolt flashes across the sky. In an ending photo montage, Henrietta has married an officer in the Grenadier Guards; David married Serena, whom he met at the second wedding; Scarlett has married Chester, the Texan at Carrie and Hamish's wedding; Tom married Deirdre; Matthew has found a new partner; Fiona is shown with Prince Charles; and Charles and Carrie have had their first child. Cast Production Writing Screenwriter Richard Curtis's own experiences as a wedding attendee inspired Four Weddings and a Funeral. According to Curtis he began writing the script at age 34, after realising he had attended 65 weddings in an 11-year period. At one wedding he was propositioned by a fellow guest, but he turned her down and forever regretted it; accordingly he based the origin of Charles and Carrie's romance on that situation. It took Curtis 17 drafts to reach the final version. He has commented on director Mike Newell's influence; "I come from a school where making it funny is what matters. Mike was obsessed with keeping it real. Every character, no matter how small, has a story, not just three funny lines. It's a romantic film about love and friendship that swims in a sea of jokes." Curtis chose to omit any mention of the characters' careers, because he didn't think a group of friends would realistically discuss their jobs while together at a wedding. Casting Curtis, Newell and the producers began the casting process for Four Weddings in early 1992. Alex Jennings was cast as Charles, but funding for the production fell through in mid-1992. Jennings would eventually go on to play a supporting role in Mindy Kaling's 2019 television miniseries adaptation of the film. The team continued holding auditions for over a year, seeing roughly 70 actors for the role of Charles before Hugh Grant. Grant was ready to give up acting as a career when he received the script for Four Weddings and a Funeral; he stated in 2016 that: "I wasn't really getting any work at all, and then to my great surprise this script came through the letterbox from my agent, and it was really good. And I rang on and said there must be a mistake, you've sent me a good script." Initially, writer Richard Curtis, who had modelled the character of Charles after himself, was opposed to casting Grant in the role, because he thought Grant was too handsome. Curtis favoured casting Alan Rickman, but Rickman refused to audition. Curtis was eventually persuaded by Newell and the producers to approve Grant's casting. Jeanne Tripplehorn was originally cast as Carrie, but she had to drop out before filming when her mother passed away. The role was offered to Marisa Tomei, but she turned it down, because her grandfather was sick at the time. Sarah Jessica Parker was also reportedly considered. Andie MacDowell was in London doing publicity for Groundhog Day when she read the script and was subsequently cast. MacDowell took a 75% cut in her fee to appear, receiving $250,000 upfront, but due to the success of the film, she earned around $3 million. Grant's participation hit another stumbling block when his agent requested a £5,000 rise over the £35,000 salary Grant was offered. The producers initially refused because of the extremely tight budget, but eventually agreed. The supporting cast-members were paid £17,500 apiece. Production Duncan Kenworthy produced the film while on sabbatical from Jim Henson Productions. Pre-production for the movie was a long process because funding was erratic, falling through in mid-1992 and leading to much uncertainty. Finally in early 1993, Working Title Films stepped in to close the gap. Nonetheless, another $1.2 million was cut just before production began in the summer of 1993, forcing the film to be made in just 36 days with a final budget of £2.7 million (appr. $4.4 million in 1994). Channel Four Films contributed £800,000. The budget was so tight that extras had to wear their own wedding clothes, while Rowan Atkinson appeared as a vicar at two of the weddings so production wouldn't have to pay another actor. Future Home Secretary and Member of Parliament (MP) Amber Rudd was given the credit of "Aristocracy Coordinator" after she arranged for several aristocrats to make uncredited appearances as wedding extras, including Peregrine Cavendish, who was at the time Marquess of Hartington, and the Earl of Woolton, who conveniently wore their own morning suits. To make Grant look more nerdy, the producers styled him with shaggy hair, glasses, and deliberately unflattering, ill-fitting clothes. Grant was encouraged by director Mike Newell to mess up and trip over his lines, written in "convoluted syntax" as Grant describes them, in order to give Charles a stammering, nervous quality. Grant, who struggled with hay fever throughout filming, was unsure of Newell's direction and his own performance, which he thought was "atrocious." Regarding Newell, Grant commented that: "He seemed to be giving direction against what I thought were the natural beats of the comedy. He was making a film with texture, grounding it, playing the truths rather than the gags". The film was shot mainly in London and the Home Counties, including: Hampstead, Islington where the final moments take place on Highbury Terrace, Greenwich Hospital, Betchworth in Surrey, Amersham in Buckinghamshire, St Bartholomew-the-Great (wedding number four) and West Thurrock in Essex. Exterior shots of guests arriving for the funeral were filmed in Thurrock, Essex overlooking the River Thames with the backdrop of the Dartford River Crossing. Stately homes in Bedfordshire (Luton Hoo for wedding two's reception) and Hampshire provided exteriors for weddings. Post-production According to Hugh Grant, the initial screening of a rough-cut of Four Weddings went very badly. Throughout production, Gramercy Pictures, the U.S. distributor for the film, sent frequent transatlantic faxes objecting to the explicit language and sexual content, fearing the final product would not be suitable for American distribution or television airings. They particularly objected to the opening scene of the movie, in which Charles and Scarlett say the word "Fuck" over and over, after an initial screening of the movie in Salt Lake City led the conservative Mormon members of the city council to walk out. Accordingly, Mike Newell and the actors agreed to reshoot the scene with the British swear word "Bugger" to be used in the American version. The executives also objected to the title, believing Four Weddings and a Funeral would turn off male viewers from the film. In its place they suggested such titles as True Love and Near Misses, Loitering in Sacred Places, Skulking Around, and Rolling in the Aisles, none of which were accepted. Music and soundtrack The original score was composed by British composer Richard Rodney Bennett. The movie also featured a soundtrack of popular songs, including a cover version of The Troggs' "Love Is All Around" performed by Wet Wet Wet that remained at number 1 on the UK Singles Chart for fifteen weeks and was then the ninth (now twelfth) biggest selling single of all time in Britain. This song would later be adapted into "Christmas Is All Around" and sung by the character of Billy Mack in Richard Curtis' 2003 film Love Actually, in which Grant also stars. The soundtrack album sold more than 750,000 units. Release Four Weddings and a Funeral had its world premiere in January 1994 at the Sundance Film Festival in Salt Lake City, Utah. It opened in the United States on 11 March 1994 in five theatres. The box office receipts from the first five days of the film's general release in the United States so impressed the movie's distributor that it decided to spend lavishly on promotion, buying full-page newspaper ads and TV-spots totaling some $11 million. The movie also benefited from much free publicity because of Grant's reception in the United States, where he became an instant sex symbol and undertook a successful media tour promoting the film. Producer Duncan Kenworthy stated that "It was the most amazing luck that when Hugh went on the publicity trail he turned out to be incredibly funny, and very like the character of Charles. That doesn't ever happen." The film had a wide release in the United States on 15 April 1994. At the UK premiere in Leicester Square on 11 May 1994, Hugh Grant's then-girlfriend Elizabeth Hurley garnered much publicity for the film when she wore a black Versace safety-pin dress which became a sensation in the press. The film opened in the UK on 13 May 1994. Reception Critical response On review aggregator Rotten Tomatoes the film holds an approval rating of 96% based on 69 reviews, with an average rating of 7.7/10. The site's critics consensus states, "While frothy to a fault, Four Weddings and a Funeral features irresistibly breezy humor, and winsome performances from Hugh Grant and Andie MacDowell." Metacritic assigned the film a weighted average score of 81 out of 100 based on 19 critics, indicating "universal acclaim". Film critic Roger Ebert gave the film three-and-a-half stars out of four, calling it "delightful and sly", and directed with "light-hearted enchantment" by Newell. He praised Grant's performance, describing it as a kind of "endearing awkwardness". Todd McCarthy of Variety called it a "truly beguiling romantic comedy" which was "frequently hilarious without being sappily sentimental or tiresomely retrograde." Producer Duncan Kenworthy later attributed much of the success of Four Weddings at the box office to McCarthy's review. Writing for the Chicago Reader, Jonathan Rosenbaum called the film "generic" and "standard issue", stating that the audience shouldn't "expect to remember it ten minutes later". Time magazine writer Richard Corliss was less scathing, but agreed that it was forgettable, saying that people would "forget all about [the movie] by the time they leave the multiplex," even joking at the end of his review that he had forgotten the film's name. Box office Upon its limited release in the United States, Four Weddings and a Funeral opened with $138,486 from five theatres. In its wide release, the film topped the box office with $4.2 million. The film would go on to gross $52.7 million in the United States and Canada. In the United Kingdom, the film grossed £1.4 million in its opening weekend, a record for a UK production, and £2.7 million in its opening week from 211 theatres. It was number one for nine consecutive weeks, grossing £27.8 million, making it the second highest-grossing film of all-time in the United Kingdom behind Jurassic Park. It surpassed A Fish Called Wanda as the highest-grossing British film. In France, it was number one at the box office for ten weeks, grossing $34.4 million. It was also number one at the Australian box office for five weeks and was the second-highest-grossing film of the year, grossing $A21.4 million. Overall, it grossed $245.7 million worldwide, generating the highest percentage return on cost of films released in 1994. The success of the film cleared Working Title's past losses and generated over $50 million for Polygram, clearing most of their losses in the four years since they started producing films. Recognition The film was voted the 27th greatest comedy film of all time by readers of Total Film in 2000. In 2004, the same magazine named it the 34th greatest British film of all time. It is number 96 on Bravo's "100 Funniest Movies". The Guardian, in a 20th anniversary retrospective of Four Weddings, stated that "Its influence on the British film industry, on romantic-comedy writing, on the pop charts, on funeral readings, on haircuts, was enormous." Hugh Grant commented in 2016 on the experience of the film's phenomenal success and its effect on his career: "I was making An Awfully Big Adventure at the time that Four Weddings came out, with Mike Newell again, same director, even tinier budget, in Dublin. And we'd get back from brutal days on the set, very long and no money, and the fax machines...were coming out saying that now your film Four Weddings is #5 in America, now it's #3, now it's #1 and here's an offer Hugh, for Captain Blood and they'll pay you $1 million. It was completely surreal." Awards and accolades Year-end lists 1st – Glenn Lovell, San Jose Mercury News 2nd – Sandi Davis, The Oklahoman 3rd – National Board of Review 5th – Joan Vadeboncoeur, Syracuse Herald American 5th – John Hurley, Staten Island Advance 6th – Peter Travers, Rolling Stone 6th – Sean P. Means, The Salt Lake Tribune 7th – Michael MacCambridge, Austin American-Statesman 7th – Kenneth Turan, Los Angeles Times 7th – Janet Maslin, The New York Times 7th – Todd Anthony, Miami New Times 7th – Steve Persall, St. Petersburg Times 8th – James Berardinelli, ReelViews 8th – Mack Bates, The Milwaukee Journal 10th – Kevin Thomas, Los Angeles Times 10th – Douglas Armstrong, The Milwaukee Journal Top 7 (not ranked) – Duane Dudek, Milwaukee Sentinel Top 9 (not ranked) – Dan Webster, The Spokesman-Review Top 10 (listed alphabetically, not ranked) – Bob Ross, The Tampa Tribune Top 10 (listed alphabetically, not ranked) – Eleanor Ringel, The Atlanta Journal-Constitution Top 10 (not ranked) – Howie Movshovitz, The Denver Post Top 10 (not ranked) – George Meyer, The Ledger Top 10 (not ranked) – Bob Carlton, The Birmingham News Best "sleepers" (not ranked) – Dennis King, Tulsa World Honorable mention – Betsy Pickle, Knoxville News-Sentinel Honorable mention – William Arnold, Seattle Post-Intelligencer Honorable mention – David Elliott, The San Diego Union-Tribune Honorable mention – Robert Denerstein, Rocky Mountain News Honorable mention – Michael Mills, The Palm Beach Post Honorable mention – Jeff Simon, The Buffalo News Awards Franchise Hulu anthology television miniseries It was reported in November 2017 that the streaming service Hulu was developing an eponymous anthology television series based upon the film, to be written and executive produced by Mindy Kaling and Matt Warburton, with Richard Curtis also serving as an executive producer. In October 2018, it was announced Jessica Williams, Nikesh Patel, Rebecca Rittenhouse, and John Reynolds had joined the cast. The miniseries premiered on 31 July 2019. One Red Nose Day and a Wedding On 5 December 2018, it was announced that Richard Curtis had written One Red Nose Day and a Wedding, a 25th anniversary Comic Relief television reunion short film. The original film's director, Mike Newell, returned, along with the film's surviving cast, including Hugh Grant, Andie MacDowell, Kristin Scott Thomas, John Hannah, Rowan Atkinson, James Fleet, David Haig, Sophie Thompson, David Bower, Robin McCaffrey, Anna Chancellor, Rupert Vansittart, Simon Kunz, Sara Crowe and Timothy Walker. It was filmed on 13–14 December 2018 at St James' Church, Islington, London. It centered on the reunion of all the characters from the original film at the wedding of Charles and Carrie's daughter to Fiona's daughter. The involvement of additional cast members Lily James and Alicia Vikander was not announced until the day the film aired in the UK, because they played the young women getting married. The film aired in the US on their Red Nose Day on Thursday 23 May 2019. See also BFI Top 100 British films Notting Hill (1999), also written by Curtis and starring Grant Love Actually (2003), another film by Curtis starring Grant and Atkinson Black Versace dress of Elizabeth Hurley, worn by Hurley to the film's premiere List of films featuring the deaf and hard of hearing Parey Hut Love References External links 1994 films 1994 comedy-drama films 1994 LGBT-related films 1994 romantic comedy films 1994 romantic drama films 1990s romantic comedy-drama films BAFTA winners (films) Best Film BAFTA Award winners Best Foreign Film César Award winners British LGBT-related films British romantic comedy-drama films British Sign Language films Film4 Productions films Films about funerals Films about weddings in the United Kingdom Films directed by Mike Newell 1994 independent films 1990s British films Working Title Films films Films featuring a Best Musical or Comedy Actor Golden Globe winning performance 1990s English-language films Films scored by Richard Rodney Bennett Films set in London Films set in Scotland Films set in Somerset Films shot at Shepperton Studios Films shot in Bedfordshire Films shot in Buckinghamshire Films shot in Essex Films shot in Hampshire Films shot in London Films whose director won the Best Direction BAFTA Award Films with screenplays by Richard Curtis Gay-related films LGBT-related romantic comedy-drama films PolyGram Filmed Entertainment films
11371
https://en.wikipedia.org/wiki/Quintus%20Fabius%20Maximus%20Verrucosus
Quintus Fabius Maximus Verrucosus
Quintus Fabius Maximus Verrucosus, surnamed Cunctator ( 280 – 203 BC), was a Roman statesman and general of the third century BC. He was consul five times (233, 228, 215, 214, and 209 BC) and was appointed dictator in 221 and 217 BC. He was censor in 230 BC. His agnomen, Cunctator, usually translated as "the delayer", refers to the strategy that he employed against Hannibal's forces during the Second Punic War. Facing an outstanding commander with superior numbers, he pursued a then-novel strategy of targeting the enemy's supply lines, and accepting only smaller engagements on favourable ground, rather than risking his entire army on direct confrontation with Hannibal himself. As a result, he is regarded as the originator of many tactics used in guerrilla warfare. Beginnings Born at Rome c. 280 BC, Fabius was a descendant of the ancient patrician Fabia gens. He was the son or grandson of Quintus Fabius Maximus Gurges, three times consul and princeps senatus, and grandson or great-grandson of Quintus Fabius Maximus Rullianus, a hero of the Samnite Wars, who like Verrucosus held five consulships, as well as the offices of dictator and censor. Many earlier ancestors had also been consuls. His cognomen, Verrucosus, or "warty", used to distinguish him from other members of his family, derived from a wart on his upper lip. According to Plutarch, Fabius possessed a mild temper and slowness in speaking. As a child, he had difficulties in learning, engaged in sports with other children cautiously and appeared submissive in his interactions with others. All the above were perceived by those who knew him superficially to be signs of inferiority. However, according to Plutarch, these traits proceeded from stability, greatness of mind, and lion-likeness of temper. By the time he reached adulthood and was roused by active life, his virtues exerted themselves; consequently, his lack of energy displayed during his earlier years was revealed as a result of a lack of passion and his slowness was recognised as a sign of prudence and firmness. While still a youth in 265 BC, Fabius was consecrated an augur. It is unknown whether he participated in the First Punic War, fought between the Roman Republic and Carthage from 264 to 241 BC, or what his role might have been. Fabius' political career began in the years following that war. He was probably quaestor in 237 or 236 BC, and curule aedile about 235. During his first consulship, in 233 BC, Fabius was awarded a triumph for his victory over the Ligurians, whom he defeated and drove into the Alps. He was censor in 230, then consul a second time in 228. It is possible that he held the office of dictator for a first time around this time: according to Livy, Fabius's tenure of the dictatorship in 217 was his second term in that office, with Gaius Flaminius as his deputy and magister equitum during the first term: however Plutarch suggests that Flaminius was deputy instead to Marcus Minucius Rufus – presumably Fabius's great political rival of that name, who later served as deputy to Fabius himself (see below). It is of course possible that Flaminius was successively deputy to both, after Minucius's apparently premature deposition following bad augural omens: and also possible that little of note (other than, possibly, holding elections during the absence of consuls) was accomplished during either dictatorship. According to Livy, in 218 BC Fabius took part in an embassy to Carthage, sent to demand redress for the capture of the supposedly neutral town of Saguntum in Spain. After the delegation had received the Carthaginians' reply, it was Fabius himself who, addressing the Carthaginian senate, issued a formal declaration of war between Carthage and the Roman Republic. However, Cassius Dio, followed by Zonaras, calls the ambassador Marcus Fabius, suggesting that it was his cousin, Marcus Fabius Buteo, who issued the declaration of war against the Carthaginians. Dictatorship during the Second Punic War When the Consul Tiberius Sempronius Longus was defeated in the Battle of the Trebia in December 218 BC Fabius advised that the Romans should simply bide their time and deny Hannibal any chance at a general engagement instead letting the invasion peter out while making sure the cities of their Italian Allies were supported or protected. However, Consul Gaius Flaminius opposed this and joined his colleague Gnaeus Servilius Geminus in raising two consular armies to confront Hannibal in central Italy. Flaminius' plan came to a disastrous end when he was killed during the decisive Roman defeat at the Battle of Lake Trasimene in 217 BC, with panic swept Rome. With Consular armies destroyed in two major battles, and Hannibal approaching Rome's gates, the Romans feared the imminent destruction of their city. The Roman Senate decided to appoint a dictator, and chose Fabius for the role – possibly for the second time, though evidence of a previous term seems to be conflicting – in part due to his advanced age and experience. However, he was not allowed to appoint his own magister equitum; instead, the Romans chose a political enemy, Marcus Minucius. Then Fabius quickly sought to calm the Roman people by asserting himself as a strong Dictator at the moment of what was perceived to be the worst crisis in Roman history. He asked the Senate to allow him to ride on horseback, which Dictators were never allowed to do. He then caused himself to be accompanied by the full complement of twenty-four lictors, and ordered the surviving Consul, Gnaeus Servilius Geminus, to dismiss his lictors (in essence, acknowledging the seniority of the dictator), and to present himself before Fabius as a private citizen. Plutarch tells us that Fabius believed that the disaster at Lake Trasimene was due, in part, to the fact that the gods had become neglected. Before that battle, a series of omens had been witnessed, including a series of lightning bolts, which Fabius had believed were warnings from the gods. He had warned Flaminius of this, but Flaminius had ignored the warnings. And so Fabius, as Dictator, next sought to please the gods. He ordered a massive sacrifice of the whole product of the next harvest season throughout Italy, in particular that of cows, goats, swine, and sheep. In addition, he ordered that musical festivities be celebrated, and then told his fellow citizens to each spend a precise sum of 333 sestertii and 333 denarii. Plutarch isn't sure exactly how Fabius came up with this number, although he believes it was to honor the perfection of the number three, as it is the first of the odd numbers, and one of the first of the prime numbers. It is not known if Fabius truly believed that these actions had won the gods over to the Roman side, although the actions probably did (as intended) convince the average Roman that the gods had finally been won over. Fabian strategy Fabius respected Hannibal's military genius and so refused to meet him in a pitched battle. Instead, he kept his troops close to Hannibal, hoping to exhaust him in a long war of attrition. Fabius was able to harass the Carthaginian foraging parties, limiting Hannibal's ability to wreak destruction, while conserving his own military force. He avoided directly engaging Hannibal, while implementing a "scorched earth" practice to prevent Hannibal's forces from obtaining grain and other resources. The Romans were unimpressed with this defensive strategy and at first gave Fabius his epithet Cunctator (delayer) as an insult. The strategy was in part ruined because of a lack of unity in the command of the Roman army, since Fabius' Master of the Horse, Minucius, was a political enemy of Fabius. At one point, Fabius was called by the priests to assist with certain sacrifices, so Fabius left the command of the army in the hands of Minucius during his absence. Fabius had told Minucius not to attack Hannibal in his absence, but Minucius disobeyed and attacked anyway. The attack, though of no strategic value, resulted in the retreat of several enemy units, and so the Roman people, desperate for good news, believed Minucius to be a hero. On hearing of this, Fabius became enraged, and as Dictator, could have ordered Minucius' execution for his disobedience. One of the plebeian tribunes (chief representatives of the people) for the year, Metilius, was a partisan of Minucius, and as such he sought to use his power to help Minucius. The Plebeian Tribunes were the only magistrates independent of the Dictator, and so with his protection, Minucius was relatively safe. Plutarch states that Metilius "boldly applied himself to the people in the behalf of Minucius", and had Minucius granted powers equivalent to those of Fabius. By this, Plutarch probably means that as a Plebeian Tribune, Metilius had the Plebeian Council, a popular assembly which only Tribunes could preside over, grant Minucius quasi-dictatorial powers. Fabius did not attempt to fight the promotion of Minucius, but rather decided to wait until Minucius' rashness caused him to run headlong into some disaster. He realized what would happen when Minucius was defeated in battle by Hannibal. Fabius, we are told, reminded Minucius that it was Hannibal, and not he, who was the enemy. Minucius proposed that they share the joint control of the army, with command rotating between the two every other day. Fabius rejected this, and instead let Minucius command half of the army, while he commanded the other half. Minucius openly claimed that Fabius was cowardly because he failed to confront the Carthaginian forces. Near Larinum in Samnium, Hannibal had taken up position in a town called Geronium. In the leadup to the battle of Geronium, Minucius decided to make a broad frontal attack on Hannibal's troops in the valley between Larinum and Geronium. Several thousand men were involved on either side. It appeared that the Roman troops were winning, but Hannibal had set a trap. Soon the Roman troops were being slaughtered. Upon seeing the ambush of Minucius' army, Fabius cried "O Hercules! how much sooner than I expected, though later than he seemed to desire, hath Minucius destroyed himself!" On ordering his army to join the battle and rescue their fellow Romans, Fabius exclaimed "we must make haste to rescue Minucius, who is a valiant man, and a lover of his country." Fabius rushed to his co-commander's assistance and Hannibal's forces immediately retreated. After the battle, there was some feeling that there would be conflict between Minucius and Fabius; however, the younger soldier marched his men to Fabius' encampment and is reported to have said, "My father gave me life. Today you saved my life. You are my second father. I recognize your superior abilities as a commander." When Fabius' term as Dictator ended, consular government was restored, and Gnaeus Servilius Geminus and Marcus Atilius Regulus assumed the consulship for the remainder of the year. The once-looked-down-upon tactics employed by Fabius came then to be respected. It is said, asserts Plutarch, that even Hannibal acknowledged and feared the Fabian strategy and the Roman inexhaustible manpower. After Fabius lured him away from Apulia into the Bruttian territory and then proceeded to besiege Tarentum by treachery in 209 BC, Hannibal commented, "It seems that the Romans have found another Hannibal, for we have lost Tarentum in the same way that we took it." After his dictatorship Shortly after Fabius had laid down his dictatorship, Gaius Terentius Varro and Lucius Aemilius Paullus were elected as consuls. They rallied the people through the assemblies, and won their support for Varro's plan to abandon Fabius' strategy, and engage Hannibal directly. Varro's rashness did not surprise Fabius, but when Fabius learned of the size of the army (eighty-eight thousand soldiers) that Varro had raised, he became quite concerned. Unlike the losses that had been suffered by Minucius, a major loss by Varro had the potential to kill so many soldiers that Rome might have had no further resources with which to continue the war. Fabius had warned the other consul for the year, Aemilius Paullus, to make sure that Varro remained unable to directly engage Hannibal. According to Plutarch, Paullus replied to Fabius that he feared the votes in Rome more than Hannibal's army. When word reached Rome of the disastrous Roman defeat under Varro and Paullus at the Battle of Cannae in 216 BC, the Senate and the People of Rome turned to Fabius for guidance. They had believed his strategy to be flawed before, but now they thought him to be as wise as the gods. He walked the streets of Rome, assured as to eventual Roman victory, in an attempt to comfort his fellow Romans. Without his support, the senate might have remained too frightened to even meet. He placed guards at the gates of the city to stop the frightened Romans from fleeing, and regulated mourning activities. He set times and places for this mourning, and ordered that each family perform such observances within their own private walls, and that the mourning should be complete within a month; following the completion of these mourning rituals, the entire city was purified of its blood-guilt in the deaths. Although he did not again hold the office of dictator}and indeed, it was granted to others over himhe might as well have been one unofficially at this time, because whatever measures he proposed were immediately adopted with little or no further debate. Honors and death Cunctator became an honorific title, and his delaying tactic was followed in Italy for the rest of the war. Fabius' own military success was small, aside from the reconquest of Tarentum in 209 BC. For this victory, Plutarch tells us, he was awarded a second triumph that was even more splendid than the first. When Marcus Livius Macatus, the governor of Tarentum, claimed the merit of recovering the town, Fabius rejoined, "Certainly, had you not lost it, I would have never retaken it." After serving as Dictator, he served as a Consul twice more (in 215 BC and 214 BC), and for a fifth time in 209 BC. He was also chief augur (at a very young age) and pontifex, but never pontifex maximus according to Gaius Stern (citing Livy on Fabius). The holding of seats in the two highest colleges was not repeated until either Julius Caesar or possibly Sulla. In the senate, he opposed the young and ambitious Scipio Africanus, who wanted to carry the war to Africa. Fabius continued to argue that confronting Hannibal directly was too dangerous. Scipio planned to take Roman forces to Carthage itself and force Hannibal to return to Africa to defend the city. Scipio was eventually given limited approval, despite continuous opposition from Fabius, who blocked levies and restricted Scipio's access to troops. Fabius wished to ensure that sufficient forces remained to defend Roman territory if Scipio was defeated. Fabius became gravely ill and died in 203 BC, shortly after Hannibal's army left Italy, and before the eventual Roman victory over Hannibal at the Battle of Zama won by Scipio. Part of his eulogy is preserved on a fragment, which praised his delaying strategy in his altercations with Hannibal during the Second Punic War. The inscription reads as follows: "...[as censor] he conducted the first revision of the senate membership and held committal elections in the consulship of Marcus Junius Pera and Marcus Barbula; he besieged and recaptured Tarentum and the strong-hold of Hannibal, and [obtained enormous booty?]; he won surpassing glory by his military [exploits?]." Legacy Later, he became a legendary figure and the model of a tough, courageous Roman, and was bestowed the honorific title, "The Shield of Rome" (similar to Marcus Claudius Marcellus being named the "Sword of Rome"). According to Ennius, unus homo nobis cunctando restituit rem"one man, by delaying, restored the state to us." Virgil, in the Aeneid, has Aeneas' father Anchises mention Fabius Maximus while in Hades as the greatest of the many great Fabii, quoting the same line. While Hannibal is mentioned in the company of history's greatest generals, military professionals have bestowed Fabius' name on an entire strategic doctrine known as "Fabian strategy", and George Washington has been called "the American Fabius". Mikhail Kutuzov has likewise been called "the Russian Fabius" for his strategy against Napoleon. According to its own ancient legend, the Roman princely family of Massimo descends from Fabius Maximus. See also Fabian Society, a British socialist society founded at the end of the 19th century and still active today. Their name derives from the tactics of Quintus Fabius Maximus. Gens Fabia List of ancient Romans Footnotes References Primary sources Fabius, by Plutarch Secondary material Encyclopædia Britannica, Fabius Maximus Cunctator Plutarch Makers of Rome translated by Ian Scott-Kilvert 1965, Penguin Books, London Livy The War with Hannibal translated by Aubrey de Selincourt 1974, Penguin Books, London Further reading McCall, Jeremiah (2018). Clan Fabius, Defenders of Rome: A History of the Republic's Most Illustrious Family. Barnslet: Pen and Sword. . External links 3rd-century BC births 203 BC deaths 3rd-century BC Roman augurs 3rd-century BC Roman consuls Ancient Roman dictators Characters in Book VI of the Aeneid Verrucosus, Quintus Pontifices Roman censors Roman commanders of the Second Punic War Roman patricians Roman triumphators Romans who received the grass crown Year of birth uncertain
11524
https://en.wikipedia.org/wiki/Fahrenheit
Fahrenheit
The Fahrenheit scale () is a temperature scale based on one proposed in 1724 by the physicist Daniel Gabriel Fahrenheit (1686–1736). It uses the degree Fahrenheit (symbol: °F) as the unit. Several accounts of how he originally defined his scale exist, but the original paper suggests the lower defining point, 0 °F, was established as the freezing temperature of a solution of brine made from a mixture of water, ice, and ammonium chloride (a salt). The other limit established was his best estimate of the average human body temperature, originally set at 90 °F, then 96 °F (about 2.6 °F less than the modern value due to a later redefinition of the scale). For much of the 20th century, the Fahrenheit scale was defined by two fixed points with a 180 °F separation: the temperature at which pure water freezes was defined as 32 °F and the boiling point of water was defined to be 212 °F, both at sea level and under standard atmospheric pressure. It is now formally defined using the Kelvin scale and hence ultimately by the Boltzmann constant, the Planck constant, and the second (defined as a specific number of cycles of the unperturbed ground-state hyperfine transition frequency of the caesium-133 atom.) It continues to be officially used in the United States (including its unincorporated territories), its freely associated states in the Western Pacific (Palau, the Federated States of Micronesia and the Marshall Islands), the Cayman Islands, and the former American colony of Liberia. Fahrenheit is used alongside the Celsius scale in Antigua and Barbuda and other countries which use the same meteorological service, such as Saint Kitts and Nevis, the Bahamas, and Belize. A handful of British Overseas Territories, including the Virgin Islands, Montserrat, Anguilla, and Bermuda, still use both scales. All other countries now use Celsius ("centigrade" until 1948), a scale formalized about 20 years after the Fahrenheit scale. The United Kingdom started to change from Fahrenheit to Celsius in 1962, and many people remain aware of Fahrenheit temperatures; degrees Fahrenheit are sometimes used in newspaper headlines to sensationalize heatwaves. Definition and conversion Historically, on the Fahrenheit scale the freezing point of water was 32 °F, and the boiling point was 212 °F (at standard atmospheric pressure). This put the boiling and freezing points of water 180 degrees apart. Therefore, a degree on the Fahrenheit scale was of the interval between the freezing point and the boiling point. On the Celsius scale, the freezing and boiling points of water were originally defined to be 100 degrees apart. A temperature interval of 1 °F was equal to an interval of  degrees Celsius. With the Fahrenheit and Celsius scales now both defined by the kelvin, this relationship was preserved, a temperature interval of 1 °F being equal to an interval of  K and of  °C. The Fahrenheit and Celsius scales intersect numerically at −40 in the respective unit (i.e, −40 °F ≘ −40 °C). Absolute zero is 0 K, −273.15 °C, or −459.67 °F. The Rankine temperature scale uses degree intervals of the same size as those of the Fahrenheit scale, except that absolute zero is 0 °R the same way that the Kelvin temperature scale matches the Celsius scale, except that absolute zero is 0 K. The combination of degree symbol (°) followed by an uppercase letter F is the conventional symbol for the Fahrenheit temperature scale. A number followed by this symbol (and separated from it with a space) denotes a specific temperature point (e.g., "Gallium melts at 85.5763 °F"). A difference between temperatures or an uncertainty in temperature is also conventionally written the same way as well, e.g., "The output of the heat exchanger experiences an increase of 72 °F" or "Our standard uncertainty is ±5 °F". However, some authors instead use the notation "An increase of " (reversing the symbol order) to indicate temperature differences. Similar conventions exist for the Celsius scale, see . Conversion (specific temperature point) For an exact conversion between degrees Fahrenheit and Celsius, and kelvins of a specific temperature point, the following formulas can be applied. Here, is the value in degrees Fahrenheit, the value in degrees Celsius, and the value in kelvins:  °F to  °C: =  °C to  °F: = × 1.8 + 32  °F to  K: =  K to  °F: = × 1.8 − 459.67 There is also an exact conversion between Celsius and Fahrenheit scales making use of the correspondence −40 °F ≘ −40 °C. Again, is the numeric value in degrees Fahrenheit, and the numeric value in degrees Celsius:  °F to  °C: = − 40  °C to  °F: = ( + 40) × 1.8 − 40 Conversion (temperature difference or interval) When converting a temperature interval between the Fahrenheit and Celsius scales, only the ratio is used, without any constant (in this case, the interval has the same numeric value in kelvins as in degrees Celsius):  °F to  °C or  K: = =  °C or  K to  °F: = × 1.8 = × 1.8 History Fahrenheit proposed his temperature scale in 1724, basing it on two reference points of temperature. In his initial scale (which is not the final Fahrenheit scale), the zero point was determined by placing the thermometer in "a mixture of ice, water, and salis Armoniaci [transl. ammonium chloride] or even sea salt". This combination forms a eutectic system, which stabilizes its temperature automatically: 0 °F was defined to be that stable temperature. A second point, 96 degrees, was approximately the human body's temperature. A third point, 32 degrees, was marked as being the temperature of ice and water "without the aforementioned salts". According to a German story, Fahrenheit actually chose the lowest air temperature measured in his hometown Danzig (Gdańsk, Poland) in winter 1708–09 as 0 °F, and only later had the need to be able to make this value reproducible using brine. According to a letter Fahrenheit wrote to his friend Herman Boerhaave, his scale was built on the work of Ole Rømer, whom he had met earlier. In Rømer scale, brine freezes at zero, water freezes and melts at 7.5 degrees, body temperature is 22.5, and water boils at 60 degrees. Fahrenheit multiplied each value by 4 in order to eliminate fractions and make the scale more fine-grained. He then re-calibrated his scale using the melting point of ice and normal human body temperature (which were at 30 and 90 degrees); he adjusted the scale so that the melting point of ice would be 32 degrees, and body temperature 96 degrees, so that 64 intervals would separate the two, allowing him to mark degree lines on his instruments by simply bisecting the interval 6 times (since 64 = 26). Fahrenheit soon after observed that water boils at about 212 degrees using this scale. The use of the freezing and boiling points of water as thermometer fixed reference points became popular following the work of Anders Celsius, and these fixed points were adopted by a committee of the Royal Society led by Henry Cavendish in 1776–77. Under this system, the Fahrenheit scale is redefined slightly so that the freezing point of water was exactly 32 °F, and the boiling point was exactly 212 °F, or 180 degrees higher. It is for this reason that normal human body temperature is approximately 98.6 °F (oral temperature) on the revised scale (whereas it was 90° on Fahrenheit's multiplication of Rømer, and 96° on his original scale). In the present-day Fahrenheit scale, 0 °F no longer corresponds to the eutectic temperature of ammonium chloride brine as described above. Instead, that eutectic is at approximately 4 °F on the final Fahrenheit scale. The Rankine temperature scale was based upon the Fahrenheit temperature scale, with its zero representing absolute zero instead. Usage General The Fahrenheit scale was the primary temperature standard for climatic, industrial and medical purposes in English-speaking countries until the 1960s. In the late 1960s and 1970s, the Celsius scale replaced Fahrenheit in almost all of those countries—with the notable exception of the United States and in certain cases, the United Kingdom—typically during their general metrication process. Fahrenheit is used in the United States, its territories and associated states (all served by the U.S. National Weather Service), as well as the Cayman Islands and Liberia for everyday applications. For example, U.S. weather forecasts, food cooking, and freezing temperatures are typically given in degrees Fahrenheit. Scientists, including meteorologists, use degrees Celsius or kelvin in all countries. United States Early in the 20th century, Halsey and Dale suggested that reasons for resistance to use the centigrade (now Celsius) system in the U.S. included the larger size of each degree Celsius and the lower zero point in the Fahrenheit system; put another way, the Fahrenheit scale is more intuitive than Celsius for describing outdoor temperatures in temperate latitudes, with 100 °F being a hot summer day and 0 °F a cold winter day. Canada Canada has passed legislation favoring the International System of Units, while also maintaining legal definitions for traditional Canadian imperial units. Canadian weather reports are conveyed using degrees Celsius with occasional reference to Fahrenheit especially for cross-border broadcasts. Fahrenheit is still used on virtually all Canadian ovens. Thermometers, both digital and analog, sold in Canada usually employ both the Celsius and Fahrenheit scales. European Union In the European Union, it is mandatory to use kelvins or degrees Celsius when quoting temperature for "economic, public health, public safety and administrative" purposes, though degrees Fahrenheit may be used alongside degrees Celsius as a supplementary unit. For example, the laundry symbols used in the United Kingdom follow the recommendations of ISO 3758:2005 showing the temperature of the washing machine water in degrees Celsius only. The equivalent label in North America uses one to six dots to denote temperature with an optional temperature in degrees Celsius. United Kingdom Although Fahrenheit was the most popular scale in the United Kingdom, for many years Celsius has been the primary scale used, and it has been taught in schools since the 1970s. Within unregulated sectors, such as journalism, the use of Fahrenheit in the United Kingdom follows no fixed pattern, with degrees Fahrenheit sometimes appearing alongside degrees Celsius. The Daily Telegraph does not mention Fahrenheit on its daily weather page, while The Times also has an all-metric daily weather page but has a Celsius-to-Fahrenheit conversion table. When publishing news stories, some UK tabloids have adopted a tendency of using degrees Celsius in headlines and discussion relating to low temperatures and Fahrenheit for mid to high temperatures. In February 2006, the writer of an article in The Times suggested that the rationale was one of emphasis: "−6 °C" sounds colder than "21 °F", and "94 °F" sounds more impressive than "34 °C". Unicode representation of symbol Unicode provides the Fahrenheit symbol at code point . However, this is a compatibility character encoded for roundtrip compatibility with legacy encodings. The Unicode standard explicitly discourages the use of this character: "The sequence + is preferred over , and those two sequences should be treated as identical for searching." See also Comparison of temperature scales Degree of frost Notes References External links Daniel Gabriel Fahrenheit (Polish-born Dutch physicist) – Encyclopædia Britannica "At Auction | One of Only Three Original Fahrenheit Thermometers" Enfilade page for 2012 Christie's sale of a Fahrenheit mercury thermometer Christie's press release Customary units of measurement in the United States Imperial units Scales of temperature 1724 introductions Scales in meteorology
11702
https://en.wikipedia.org/wiki/Flirting
Flirting
Flirting or coquetry is a social and sexual behavior involving body language, or spoken or written communication. It is used to suggest interest in a deeper relationship with another person and, if done playfully, for amusement. A person will flirt with another by speaking or behaving in such a way that suggests their desire to increase intimacy in their current relationship with that person. The approach may include communicating a sense of playfulness, irony, or by using double entendres. Communication characteristics Flirting takes the form of several different styles, verbal and non-verbal forms of communication are used in flirting. Verbal A person can indicate flirting behavior by changing their verbal affect: Alterations in vocal tone, such as pace, volume, and intonation. Challenging their partner in order to increase tension and test the partner's reception to flirting. Alternative forms include teasing, asking probing questions, and feigning disinterest. Adoration of their partner, by praising the partner's self-assurance, intelligence, fashion style, etc. Body language Non-verbal flirting behavior can include: Flicking of the hair Prolonged eye contact Brief touching Open stance/posture Proximity, and other gestures. Etymology The origin of the word "flirt" is unknown. The Oxford English Dictionary (first edition) associates the verb form—first used in 1580—with the intransitive "flit" and the noun form—ca 1590—with the transitive "flick". Flirt has been attributed to the French conter fleurette, which has fallen out of common use. The French word "fleurette" (small flower) was used in the 16th century in some sonnets and texts. While this term is now seen as outdated, this expression is still used in French, often mockingly, however the English gallicism, "to flirt" is in the common vernacular and has now become an anglicism. Historical context During World War II, anthropologist Margaret Mead was working in Britain for the British Ministry of Information and later for the U.S. Office of War Information, delivering speeches and writing articles to help the American soldiers better understand the British civilians, and vice versa. She observed in the flirtations between the American soldiers and British women a pattern of misunderstandings regarding who is supposed to take which initiative. She wrote of the Americans, "The boy learns to make advances and rely upon the girl to repulse them whenever they are inappropriate to the state of feeling between the pair", as contrasted to the British, where "the girl is reared to depend upon a slight barrier of chilliness... which the boys learn to respect, and for the rest to rely upon the men to approach or advance, as warranted by the situation." When flirting with each other, British women could interpret an American soldier's gregariousness as something more intimate or serious than he had intended. Communications theorist Paul Watzlawick researched courtship behaviors between English women and North American servicemen in late- to post-WWII, finding common misunderstandings of intent. The simple act of kissing during the 'wrong stage'—according to cultural mores—of the courtship often led both parties to believe the other was being too forward, too soon. European hand fans The fan was extensively used as a means of communication and therefore a way of flirting from the 16th century onwards in some European societies, especially England and Spain. A whole sign language was developed with the use of the fan, and even etiquette books and magazines were published. Charles Francis Badini created the Original Fanology or Ladies' Conversation Fan, which was published by William Cock in London in 1797. The use of the fan was not limited to women, as men also carried fans and learned how to convey messages with them. For instance, placing the fan near the heart meant "I love you", while opening a fan wide meant "Wait for me". In Spain, where the use of fans (called "abanicos") is still very popular today, ladies used them to communicate with suitors or prospective suitors without attracting the notice of their families or chaperons. This use was highly popular during the 19th and early 20th centuries. Purpose People flirt for a variety of reasons. According to social anthropologist Kate Fox, there are two main types of flirting: flirting just for fun and flirting with further intent. In a 2014 review, sociologist David Henningsen identified six main motivations for flirting: sex, relational development, exploration, fun, self-esteem, and as a means to an end. Henningsen found that many flirting interactions involve more than one of these motives. There also appear to be gender differences in flirting motivation. Courtship Many people flirt as a courtship initiation method, with the aim of engaging in a sexual relationship with another person. In this sense, flirting plays a role in the mate-selection process. The person flirting will send out signals of sexual availability to another, and expects to see the interest returned in order to continue flirting. Flirting can involve non-verbal signs, such as an exchange of glances, hand-touching, and hair-touching; or verbal signs, such as chatting, giving flattering comments, and exchanging telephone numbers in order to initiate further contact. Many studies have confirmed that sex is a motivation for flirting. Additionally, Messman and colleagues' study provided support for this hypothesis; it demonstrated that, the more one was physically attracted to a person, the higher the chances one would flirt with them. Misinterpretation and consequences Flirting with the goal of signaling interest appears as a puzzling phenomenon when considering that flirting is often performed very subtly. In fact, evidence shows that people are often mistaken in how they interpret flirting behaviors. Logically, if the main purpose of flirting is to signal interest to the other person, then the signaling would be done clearly and explicitly. A possible explanation for the ambiguous nature of human flirting lies in the costs associated with courtship signals. Indeed, according to Gersick and colleagues, signaling interest can be costly as it can lead to the disturbance of the nature of a relationship. Third parties can impose costs on individuals expressing sexual interest. Expressing sexual interest to somebody else's romantic partner is a highly punishable act. This often leads to jealousy from the person's partner which can trigger anger and (possible) physical punishment, especially in men. Third parties can also impose costs through the act of eavesdropping. These can lead to damage to one's reputation leading to possible social, economic and legal costs. The costs associated with interest signaling are magnified in the case of humans when compared to the animal world, as the existence of language means information can circulate much faster. For instance, in the case of eavesdropping, the information overhead by the eavesdropper can be spread to very large social networks, thereby magnifying the social costs. Other motivations Another reason people engage in flirting is to consolidate or maintain a romantic relationship with their partner. They will engage in flirting behaviors to promote the flourishing of their relationship with their partner. People will also flirt with the goal of 'exploring'. In this sense, the aim is not necessarily to express sexual or romantic interest but simply to assess whether the other might be interested in them before making any decision about what they would want from that individual. Henningsen and Fox also demonstrated that flirting can sometimes be employed just for fun. For instance, studies have shown that flirting in the workplace was used mostly for fun purposes. Another motive that drives flirting is developing one's own self-esteem by encouraging reciprocation. Gender differences in motivations Certain types of flirting seem to vary by gender. Henningsen and colleagues' study demonstrated that flirting with sexual intent was found to be more prominent amongst men while flirting for relationship development purposes was more often employed by women. Additionally, Henningsen found that women may engage in what he calls "practice flirting," that is, using the behavior to evaluate potential partners. In evolutionary biology, the parental investment theory states that females are more selective and males are more competitive, therefore predicting that flirting as courtship initiation will be more commonly used by males. The theory also predicts that females provide more resources to their offspring, which causes them to invest in a mate that can contribute to their offspring's survival. Cultural variations Flirting behavior varies culture to culture due to different modes of social etiquette, such as how closely people should stand, how long to hold eye contact, how much touching is appropriate and so forth. Nonetheless, some behaviors may be more universal. Ethologist Irenäus Eibl-Eibesfeldt discovered women from different continents (Africa and North America) behave similarly in some ways when flirting, such as nonchalantly breaking their gaze and smiling after first staring for a prolonged period of time. In "contact cultures," such as those in the Mediterranean or Latin America, closer proximity is common, compared with cultures such as those in Britain or Northern Europe. The variation in social norms may lead to different interpretations of what is considered to be flirting. Japanese courtesans had another form of flirting, emphasizing non-verbal relationships by hiding the lips and showing the eyes, as depicted in much Shunga art, the most popular print media at the time, until the late 19th century. The fan was extensively used as a means of communication and therefore a way of flirting from the 16th century onwards in some European societies, especially England and Spain. A whole sign language was developed with the use of the fan, and even etiquette books and magazines were published. Charles Francis Badini created the Original Fanology or Ladies' Conversation Fan, which was published by William Cock in London in 1797. The use of the fan was not limited to women, as men also carried fans and learned how to convey messages with them. For instance, placing the fan near the heart meant "I love you", while opening a fan wide meant "Wait for me". In Spain, where the use of fans (called "abanicos") is still very popular today, ladies used them to communicate with suitors or prospective suitors without attracting the notice of their families or chaperons. This use was highly popular during the 19th and early 20th centuries. In Japan, flirting in the street or public places is known as nanpa. See also Anti-Flirt Club Making out Public display of affection Wingman (social) References Sexual attraction Philosophy of love
11921
https://en.wikipedia.org/wiki/Gambling
Gambling
Gambling (also known as betting or gaming) is the wagering of something of value ("the stakes") on a random event with the intent of winning something else of value, where instances of strategy are discounted. Gambling thus requires three elements to be present: consideration (an amount wagered), risk (chance), and a prize. The outcome of the wager is often immediate, such as a single roll of dice, a spin of a roulette wheel, or a horse crossing the finish line, but longer time frames are also common, allowing wagers on the outcome of a future sports contest or even an entire sports season. The term "gaming" in this context typically refers to instances in which the activity has been specifically permitted by law. The two words are not mutually exclusive; i.e., a "gaming" company offers (legal) "gambling" activities to the public and may be regulated by one of many gaming control boards, for example, the Nevada Gaming Control Board. However, this distinction is not universally observed in the English-speaking world. For instance, in the United Kingdom, the regulator of gambling activities is called the Gambling Commission (not the Gaming Commission). The word gaming is used more frequently since the rise of computer and video games to describe activities that do not necessarily involve wagering, especially online gaming, with the new usage still not having displaced the old usage as the primary definition in common dictionaries. "Gaming" has also been used to circumvent laws against "gambling". The media and others have used one term or the other to frame conversations around the subjects, resulting in a shift of perceptions among their audiences. Gambling is also a major international commercial activity, with the legal gambling market totaling an estimated $335 billion in 2009. In other forms, gambling can be conducted with materials that have a value, but are not real money. For example, players of marbles games might wager marbles, and likewise games of Pogs or Magic: The Gathering can be played with the collectible game pieces (respectively, small discs and trading cards) as stakes, resulting in a meta-game regarding the value of a player's collection of pieces. History Gambling dates back to the Paleolithic period, before written history. In Mesopotamia the earliest six-sided dice date to about 3000 BCE. However, they were based on astragali dating back thousands of years earlier. In China, gambling houses were widespread in the first millennium BCE, and betting on fighting animals was common. Lotto games and dominoes (precursors of Pai Gow) appeared in China as early as the 10th century. Playing cards appeared in the 9th century CE in China. Records trace gambling in Japan back at least as far as the 14th century. Poker, the most popular U.S. card game associated with gambling, derives from the Persian game As-Nas, dating back to the 17th century. The first known casino, the Ridotto, started operating in 1638 in Venice, Italy. Great Britain Gambling has been a main recreational activity in Great Britain for centuries. Queen Elizabeth I chartered a lottery that was drawn in 1569. Horseracing has been a favorite theme for over three centuries. It has been heavily regulated. Historically much of the opposition comes from Nonconformist Protestants, and from social reformers. United States Gambling has been a popular activity in the United States for centuries. It has also been suppressed by law in many areas for almost as long. By the early 20th century, gambling was almost uniformly outlawed throughout the U.S. and thus became a largely illegal activity, helping to spur the growth of the mafia and other criminal organizations. The late 20th century saw a softening in attitudes towards gambling and a relaxation of laws against it. Regulation Many jurisdictions, local as well as national, either ban gambling or heavily control it by licensing the vendors. Such regulation generally leads to gambling tourism and illegal gambling in the areas where it is not allowed. The involvement of governments, through regulation and taxation, has led to a close connection between many governments and gambling organizations, where legal gambling provides significant government revenue, such as in Monaco and Macau, China. There is generally legislation requiring that gambling devices be statistically random, to prevent manufacturers from making some high-payoff results impossible. Since these high payoffs have very low probability, a house bias can quite easily be missed unless the devices are checked carefully. Most jurisdictions that allow gambling require participants to be above a certain age. In some jurisdictions, the gambling age differs depending on the type of gambling. For example, in many American states one must be over 21 to enter a casino, but may buy a lottery ticket after turning 18. Insurance Because contracts of insurance have many features in common with wagers, insurance contracts are often distinguished in law as agreements in which either party has an interest in the "bet-upon" outcome beyond the specific financial terms. e.g.: a "bet" with an insurer on whether one's house will burn down is not gambling, but rather insurance – as the homeowner has an obvious interest in the continued existence of their home independent of the purely financial aspects of the "bet" (i.e. the insurance policy). Nonetheless, both insurance and gambling contracts are typically considered aleatory contracts under most legal systems, though they are subject to different types of regulation. Asset recovery Under common law, particularly English Law (English unjust enrichment), a gambling contract may not give a casino bona fide purchaser status, permitting the recovery of stolen funds in some situations. In Lipkin Gorman v Karpnale Ltd, where a solicitor used stolen funds to gamble at a casino, the House of Lords overruled the High Court's previous verdict, adjudicating that the casino return the stolen funds less those subject to any change of position defence. U.S. Law precedents are somewhat similar. For case law on recovery of gambling losses where the loser had stolen the funds see "Rights of owner of stolen money as against one who won it in gambling transaction from thief". An interesting question is what happens when the person trying to make recovery is the gambler's spouse, and the money or property lost was either the spouse's, or was community property. This was a minor plot point in a Perry Mason novel, The Case of the Singing Skirt, and it cites an actual case Novo v. Hotel Del Rio. Religious views Buddhism Lord Buddha stated gambling as a source of destruction in Singalovada Sutra. Hinduism Ancient Hindu poems like the Gambler's Lament and the Mahabharata testify to the existence of gambling among ancient Indians, while highlighting its destructive impact. The text Arthashastra ( BCE) recommends taxation and control of gambling. Judaism Ancient Jewish authorities frowned on gambling, even disqualifying professional gamblers from testifying in court. Christianity Catholicism The Catholic Church holds the position that there is no moral impediment to gambling, so long as it is fair, all bettors have a reasonable chance of winning, there is no fraud involved, and the parties involved do not have actual knowledge of the outcome of the bet (unless they have disclosed this knowledge), and as long as the following conditions are met: the gambler can afford to lose the bet, and stops when the limit is reached, and the motivation is entertainment and not personal gain leading to the "love of money" or making a living. In general, Catholic bishops have opposed casino gambling on the grounds that it too often tempts people into problem gambling or addiction, and has particularly negative effects on poor people; they sometimes also cite secondary effects such as increases in loan sharking, prostitution, corruption, and general public immorality. Some parish pastors have also opposed casinos for the additional reason that they would take customers away from church bingo and annual festivals where games such as blackjack, roulette, craps, and poker are used for fundraising. St. Thomas Aquinas wrote that gambling should be especially forbidden where the losing bettor is underage or otherwise not able to consent to the transaction. Gambling has often been seen as having social consequences, as satirized by Balzac. For these social and religious reasons, most legal jurisdictions limit gambling, as advocated by Pascal. Protestantism Gambling views among Protestants vary, with some either discouraging or forbidding their members from participation in gambling. Methodists, in accordance with the doctrine of outward holiness, oppose gambling which they believe is a sin that feeds on greed. Other denominations that discourage gambling are the United Methodist Church, the Free Methodist Church, the Evangelical Wesleyan Church, the Salvation Army, and the Church of the Nazarene. Other Protestants that oppose gambling include Mennonites, Schwarzenau Brethren, Quakers, the Christian Reformed Church in North America, the Church of the Lutheran Confession, the Southern Baptist Convention, the Assemblies of God, and the Seventh-day Adventist Church. Other Christian denominations Other churches that oppose gambling include the Jehovah's Witnesses, The Church of Jesus Christ of Latter-day Saints, the Iglesia ni Cristo, and the Members Church of God International. Islam There is a consensus among the ‘Ulema’ (, Scholars (of Islam)) that gambling is haraam (, sinful or forbidden). In assertions made during its prohibition, Muslim jurists describe gambling as being both un-Qur’anic, and as being generally harmful to the Muslim Ummah (, Community). The Arabic terminology for gambling is Maisir. In parts of the world that implement full Shari‘ah, such as Aceh, punishments for Muslim gamblers can range up to 12 lashes or a one-year prison term and a fine for those who provide a venue for such practises. Some Islamic nations prohibit gambling; most other countries regulate it. Bahá'í Faith According to the Most Holy Book, paragraph 155, gambling is forbidden. Types Casino games While almost any game can be played for money, and any game typically played for money can also be played just for fun, some games are generally offered in a casino setting. Table games Electronic gambling Online roulette Pachinko Sic Bo Slot machine Video poker Video bingo Other gambling Bingo Keno Non-casino games Gambling games that take place outside of casinos include bingo (as played in the US and UK), dead pool, lotteries, pull-tab games and scratchcards, and Mahjong. Other non-casino gambling games include: Non-casino card games, including historical games like Basset, Ecarté, Lansquenet and Put. Technically, a gambling card game is one in which the cards are not actually played but simply bet on. Carnival Games such as The Razzle or Hanky Pank Coin-tossing games such as Head and Tail, Two-up* Confidence tricks such as Three-card Monte or the Shell game Dice-based games, such as Backgammon, Liar's dice, Passe-dix, Hazard, Threes, Pig, or Mexico (or Perudo); *Although coin tossing is not usually played in a casino, it has been known to be an official gambling game in some Australian casinos Fixed-odds betting Fixed-odds betting and Parimutuel betting frequently occur at many types of sporting events, and political elections. In addition many bookmakers offer fixed odds on a number of non-sports related outcomes, for example the direction and extent of movement of various financial indices, the winner of television competitions such as Big Brother, and election results. Interactive prediction markets also offer trading on these outcomes, with "shares" of results trading on an open market. Parimutuel betting One of the most widespread forms of gambling involves betting on horse or greyhound racing. Wagering may take place through parimutuel pools, or bookmakers may take bets personally. Parimutuel wagers pay off at prices determined by support in the wagering pools, while bookmakers pay off either at the odds offered at the time of accepting the bet; or at the median odds offered by track bookmakers at the time the race started. Sports betting Betting on team sports has become an important service industry in many countries. Before the advent of the internet, millions of people played the football pools every week in the United Kingdom. In addition to organized sports betting, both legal and illegal, there are many side-betting games played by casual groups of spectators, such as NCAA basketball tournament Bracket Pools, Super Bowl Squares, Fantasy Sports Leagues with monetary entry fees and winnings, and in-person spectator games like Moundball. Virtual sports Based on Sports Betting, Virtual Sports are fantasy and never played sports events made by software that can be played every time without wondering about external things like weather conditions. Arbitrage betting Arbitrage betting is a theoretically risk-free betting system in which every outcome of an event is bet upon so that a known profit will be made by the bettor upon completion of the event regardless of the outcome. Arbitrage betting is a combination of the ancient art of arbitrage trading and gambling, which has been made possible by the large numbers of bookmakers in the marketplace, creating occasional opportunities for arbitrage. Other types of betting One can also bet with another person that a statement is true or false, or that a specified event will happen (a "back bet") or will not happen (a "lay bet") within a specified time. This occurs in particular when two people have opposing but strongly held views on truth or events. Not only do the parties hope to gain from the bet, they place the bet also to demonstrate their certainty about the issue. Some means of determining the issue at stake must exist. Sometimes the amount bet remains nominal, demonstrating the outcome as one of principle rather than of financial importance. Betting exchanges allow consumers to both back and lay at odds of their choice. Similar in some ways to a stock exchange, a bettor may want to back a horse (hoping it will win) or lay a horse (hoping it will lose, effectively acting as bookmaker). Spread betting allows gamblers to wagering on the outcome of an event where the pay-off is based on the accuracy of the wager, rather than a simple "win or lose" outcome. For example, a wager can be based on the when a point is scored in the game in minutes and each minute away from the prediction increases or reduces the payout. Staking systems Many betting systems have been created in an attempt to "beat the house" but no system can make a mathematically unprofitable bet in terms of expected value profitable over time. Widely used systems include: Card counting – Many systems exist for blackjack to keep track of the ratio of ten values to all others; when this ratio is high the player has an advantage and should increase the amount of their bets. Keeping track of cards dealt confers an advantage in other games as well. Due-column betting – A variation on fixed profits betting in which the bettor sets a target profit and then calculates a bet size that will make this profit, adding any losses to the target. Fixed profits – the stakes vary based on the odds to ensure the same profit from each winning selection. Fixed stakes – a traditional system of staking the same amount on each selection. Kelly – the optimum level to bet to maximize your future median bank level. Martingale – A system based on staking enough each time to recover losses from previous bet(s) until one wins. Other uses of the term Many risk-return choices are sometimes referred to colloquially as "gambling." Whether this terminology is acceptable is a matter of debate: Emotional or physical risk-taking, where the risk-return ratio is not quantifiable (e.g., skydiving, campaigning for political office, asking someone for a date, etc.) Insurance is a method of shifting risk from one party to another. Insurers use actuarial methods to calculate appropriate premiums, which is similar to calculating gambling odds. Insurers set their premiums to obtain a long term positive expected return in the same manner that professional gamblers select which bets to make. While insurance is sometimes distinguished from gambling by the requirement of an insurable interest, the equivalent in gambling is simply betting against one's own best interests (e.g., a sports coach betting against his own team to mitigate the financial repercussions of a losing season). Situations where the possible return is of secondary importance to the wager/purchase (e.g. entering a raffle in support of a charitable cause) Investments are also usually not considered gambling, although some investments can involve significant risk. Examples of investments include stocks, bonds and real estate. Starting a business can also be considered a form of investment. Investments are generally not considered gambling when they meet the following criteria: Economic utility Positive expected returns (at least in the long term) Underlying value independent of the risk being undertaken Some speculative investment activities are particularly risky, but are sometimes perceived to be different from gambling: Foreign currency exchange (forex) transactions Prediction markets Securities derivatives, such as options or futures, where the value of the derivative is dependent on the value of the underlying asset at a specific point in time (typically the derivative's associated expiration date) Negative consequences Studies show that though many people participate in gambling as a form of recreation or to earn an income, gambling, like any behavior involving variation in brain chemistry, can become a behavioral addiction. Behavioral addiction can occur with all the negative consequences in a person's life minus the physical issues faced by people who compulsively engage in drug and alcohol abuse. Problem gambling has multiple symptoms. Gamblers often gamble to try to win back money they have lost, and some gamble to relieve feelings of helplessness and anxiety. In the United Kingdom, the Advertising Standards Authority has censured several betting firms for advertisements disguised as news articles suggesting falsely that a person had cleared debts and paid for medical expenses by gambling online . The firms face possible fines. A 2020 study of 32 countries found that the greater the amount of gambling activity in a given country, the more volatile that country's stock market prices are. Psychological biases Gamblers may exhibit a number of cognitive and motivational biases that distort the perceived odds of events and that influence their preferences for gambles. Preference for likely outcomes. When gambles are selected through a choice process – when people indicate which gamble they prefer from a set of gambles (e.g., win/lose, over/under) – people tend to prefer to bet on the outcome that is more likely to occur. Bettors tend to prefer to bet on favorites in athletic competitions, and sometimes will accept even bets on favorites when offered more favorable bets on the less likely outcome (e.g., an underdog team). Optimism/Desirability Bias. Gamblers also exhibit optimism, overestimating the likelihood that desired events will occur. Fans of NFL underdog teams, for example, will prefer to bet on their teams at even odds than to bet on the favorite, whether the bet is $5 or $50. Reluctance to bet against (hedge) desired outcomes. People are reluctant to bet against desired outcomes that are relevant to their identity. Gamblers exhibit reluctance to bet against the success of their preferred U.S. presidential candidates and Major League Baseball, National Football League, National Collegiate Athletic Association (NCAA) basketball, and NCAA hockey teams. More than 45% of NCAA fans in Studies 5 and 6, for instance, turned down a "free" real $5 bet against their team. From a psychological perspective, such a "hedge" creates an interdependence dilemma – a motivational conflict between a short-term monetary gain and the long-term benefits accrued from feelings of identification with and loyalty to a position, person, or group whom the bettor desires to succeed. In economic terms, this conflicted decision can be modeled as a trade-off between the outcome utility gained by hedging (e.g., money) and the diagnostic costs it incurs (e.g., disloyalty). People make inferences about their beliefs and identity from their behavior. If a person is uncertain about an aspect of their identity, such as the extent to which they values a candidate or team, hedging may signal to them that they are not as committed to that candidate or team as they originally believed. If the diagnostic cost of this self-signal and the resulting identity change are substantial, it may outweigh the outcome utility of hedging, and they may reject even very generous hedges. Ratio bias. Gamblers will prefer gambles with worse odds that are drawn from a large sample (e.g., drawing one red ball from an urn containing 89 red balls and 11 blue balls) to better odds that are drawn from a small sample (drawing one red ball from an urn containing 9 red balls and one blue ball). Gambler's fallacy/positive recency bias. See also References Further reading Chambers, Kerry. Gambling for profit: Lotteries, gaming machines, and casinos in cross-national focus (U of Toronto press, 2011). Ferentzy, Peter, and Nigel Turner. "Gambling and organized crime-A review of the literature." Journal of Gambling Issues 23 (2009): 111–155. Ferentzy, Peter, and Nigel E. Turner. A history of problem gambling (Springer-Verlag, 2013) Haller, Mark H. "The changing structure of American gambling in the twentieth century." Journal of Social Issues 35.3 (1979): 87–114. Richard, Brian. "Diffusion of an economic development policy innovation: Explaining the international spread of casino gambling." Journal of Gambling Studies 26.2 (2010): 287–300. Online Schwartz, David G. Roll The Bones: The History of Gambling (2006), scholarly history with global perspective excerpt External links Center for Gaming Research – at University of Nevada, Las Vegas Institute for the Study of Gambling and Commercial Gaming at the University of Nevada, Reno
12104
https://en.wikipedia.org/wiki/Guglielmo%20Marconi
Guglielmo Marconi
Guglielmo Giovanni Maria Marconi, 1st Marquis of Marconi (; 25 April 187420 July 1937) was an Italian inventor and electrical engineer, known for his creation of a practical radio wave-based wireless telegraph system. This led to Marconi being credited as the inventor of radio, and he shared the 1909 Nobel Prize in Physics with Karl Ferdinand Braun "in recognition of their contributions to the development of wireless telegraphy". Marconi was also an entrepreneur, businessman, and founder of The Wireless Telegraph & Signal Company in the United Kingdom in 1897 (which became the Marconi Company). In 1929, Marconi was ennobled as a Marchese (marquis) by King Victor Emmanuel III of Italy, and, in 1931, he set up Vatican Radio for Pope Pius XI. Biography Early years Marconi was born into the Italian nobility as Guglielmo Giovanni Maria Marconi in Palazzo Marescalchi in Bologna on 25 April 1874, the second son of Giuseppe Marconi (an Italian aristocratic landowner from Porretta Terme) and his Irish wife Annie Jameson (daughter of Andrew Jameson of Daphne Castle in County Wexford, Ireland, and granddaughter of John Jameson, founder of whiskey distillers Jameson & Sons). Marconi had a brother, Alfonso, and a stepbrother, Luigi. Between the ages of two and six, Marconi and his older brother Alfonso lived with their mother in the English town of Bedford. Education Marconi did not attend school as a child and did not go on to formal higher education. Instead, he learned chemistry, mathematics, and physics at home from a series of private tutors hired by his parents. His family hired additional tutors for Guglielmo in the winter when they would leave Bologna for the warmer climate of Tuscany or Florence. Marconi noted an important mentor was professor Vincenzo Rosa, a high school physics teacher in Livorno. Rosa taught the 17-year-old Marconi the basics of physical phenomena as well as new theories on electricity. At the age of 18 and back in Bologna, Marconi became acquainted with University of Bologna physicist Augusto Righi, who had done research on Heinrich Hertz's work. Righi permitted Marconi to attend lectures at the university and also to use the university's laboratory and library. Radio work From youth, Marconi was interested in science and electricity. In the early 1890s, he began working on the idea of "wireless telegraphy"—i.e., the transmission of telegraph messages without connecting wires as used by the electric telegraph. This was not a new idea; numerous investigators and inventors had been exploring wireless telegraph technologies and even building systems using electric conduction, electromagnetic induction and optical (light) signalling for over 50 years, but none had proven technically and commercially successful. A relatively new development came from Heinrich Hertz, who, in 1888, demonstrated that one could produce and detect electromagnetic radiation, based on the work of James Clerk Maxwell. At the time, this radiation was commonly called "Hertzian" waves, and is now generally referred to as radio waves. There was a great deal of interest in radio waves in the physics community, but this interest was in the scientific phenomenon, not in its potential as a communication method. Physicists generally looked on radio waves as an invisible form of light that could only travel along a line of sight path, limiting its range to the visual horizon like existing forms of visual signaling. Hertz's death in 1894 brought published reviews of his earlier discoveries including a demonstration on the transmission and detection of radio waves by the British physicist Oliver Lodge and an article about Hertz's work by Augusto Righi. Righi's article renewed Marconi's interest in developing a wireless telegraphy system based on radio waves, a line of inquiry that Marconi noted other inventors did not seem to be pursuing. Developing radio telegraphy thumb|Marconi's first transmitter incorporating a monopole antenna. It consisted of an elevated copper sheet (top) connected to a Righi spark gap (left) powered by an induction coil (center) with a telegraph key (right) to switch it on and off to spell out text messages in Morse code. At the age of 20, Marconi began to conduct experiments in radio waves, building much of his own equipment in the attic of his home at the Villa Griffone in Pontecchio (now an administrative subdivision of Sasso Marconi), Italy, with the help of his butler, Mignani. Marconi built on Hertz's original experiments and, at the suggestion of Righi, began using a coherer, an early detector based on the 1890 findings of French physicist Édouard Branly and used in Lodge's experiments, that changed resistance when exposed to radio waves. In the summer of 1894, he built a storm alarm made up of a battery, a coherer, and an electric bell, which went off when it picked up the radio waves generated by lightning. Late one night, in December 1894, Marconi demonstrated a radio transmitter and receiver to his mother, a set-up that made a bell ring on the other side of the room by pushing a telegraphic button on a bench. Supported by his father, Marconi continued to read through the literature and picked up on the ideas of physicists who were experimenting with radio waves. He developed devices, such as portable transmitters and receiver systems, that could work over long distances, turning what was essentially a laboratory experiment into a useful communication system. Marconi came up with a functional system with many components: A relatively simple oscillator or spark-producing radio transmitter; A wire or metal sheet capacity area suspended at a height above the ground; A coherer receiver, which was a modification of Édouard Branly's original device with refinements to increase sensitivity and reliability; A telegraph key to operate the transmitter to send short and long pulses, corresponding to the dots-and-dashes of Morse code; and A telegraph register activated by the coherer which recorded the received Morse code dots and dashes onto a roll of paper tape. In the summer of 1895, Marconi moved his experiments outdoors on his father's estate in Bologna. He tried different arrangements and shapes of antenna but even with improvements he was able to transmit signals only up to one half-mile, a distance Oliver Lodge had predicted in 1894 as the maximum transmission distance for radio waves. Transmission breakthrough A breakthrough came in the summer of 1895, when Marconi found that much greater range could be achieved after he raised the height of his antenna and, borrowing from a technique used in wired telegraphy, grounded his transmitter and receiver. With these improvements, the system was capable of transmitting signals up to and over hills. The monopole antenna reduced the frequency of the waves compared to the dipole antennas used by Hertz, and radiated vertically polarized radio waves which could travel longer distances. By this point, he concluded that a device could become capable of spanning greater distances, with additional funding and research, and would prove valuable both commercially and militarily. Marconi's experimental apparatus proved to be the first engineering-complete, commercially successful radio transmission system. Marconi applied to the Ministry of Post and Telegraphs, then under the direction of Maggiorino Ferraris, explaining his wireless telegraph machine and asking for funding, but never received a response. An apocryphal tale claims that the minister (incorrectly named first as Emilio Sineo, later as Pietro Lacava) wrote "to the Longara" on the document, referring to the insane asylum on Via della Lungara in Rome, but the letter was never found. In 1896, Marconi spoke with his family friend Carlo Gardini, Honorary Consul at the United States Consulate in Bologna, about leaving Italy to go to Great Britain. Gardini wrote a letter of introduction to the Ambassador of Italy in London, Annibale Ferrero, explaining who Marconi was and about his extraordinary discoveries. In his response, Ambassador Ferrero advised them not to reveal Marconi's results until after a patent was obtained. He also encouraged Marconi to come to Britain, where he believed it would be easier to find the necessary funds to convert his experiments into practical use. Finding little interest or appreciation for his work in Italy, Marconi travelled to London in early 1896 at the age of 21, accompanied by his mother, to seek support for his work. (He spoke fluent English in addition to Italian.) Marconi arrived at Dover, and the Customs officer opened his case to find various apparatus. The customs officer immediately contacted the Admiralty in London. While there, Marconi gained the interest and support of William Preece, the Chief Electrical Engineer of the General Post Office (the GPO). During this time Marconi decided he should patent his system, which he applied for on 2 June 1896, British Patent number 12039 titled "Improvements in Transmitting Electrical impulses and Signals, and in Apparatus therefor", which would become the first patent for a radio wave based communication system. Demonstrations and achievements Marconi made the first demonstration of his system for the British government in July 1896. A further series of demonstrations for the British followed, and, by March 1897, Marconi had transmitted Morse code signals over a distance of about across Salisbury Plain. On 13 May 1897, Marconi sent the first ever wireless communication over open sea – a message was transmitted over the Bristol Channel from Flat Holm Island to Lavernock Point near Cardiff, a distance of . The message read, "Are you ready". The transmitting equipment was almost immediately relocated to Brean Down Fort on the Somerset coast, stretching the range to . Impressed by these and other demonstrations, Preece introduced Marconi's ongoing work to the general public at two important London lectures: "Telegraphy without Wires", at the Toynbee Hall on 11 December 1896; and "Signalling through Space without Wires", given to the Royal Institution on 4 June 1897. Numerous additional demonstrations followed, and Marconi began to receive international attention. In July 1897, he carried out a series of tests at La Spezia, in his home country, for the Italian government. A test for Lloyd's between The Marine Hotel in Ballycastle and Rathlin Island, both in County Antrim in Ulster, Ireland, was conducted on 6 July 1898 by George Kemp and Edward Edwin Glanville. A transmission across the English channel was accomplished on 27 March 1899, from Wimereux, France to South Foreland Lighthouse, England. Marconi set up an experimental base at the Haven Hotel, Sandbanks, Poole Harbour, Dorset, where he erected a 100-foot high mast. He became friends with the van Raaltes, the owners of Brownsea Island in Poole Harbour, and his steam yacht, the Elettra, was often moored on Brownsea or at The Haven Hotel. Marconi purchased the vessel after the Great War and converted it to a seaborne laboratory from where he conducted many of his experiments. Among the Elettras crew was Adelmo Landini, his personal radio operator, who was also an inventor. In December 1898, the British lightship service authorised the establishment of wireless communication between the South Foreland lighthouse at Dover and the East Goodwin lightship, twelve miles distant. On 17 March 1899, the East Goodwin lightship sent the first wireless distress signal, a signal on behalf of the merchant vessel Elbe which had run aground on Goodwin Sands. The message was received by the radio operator of the South Foreland lighthouse, who summoned the aid of the Ramsgate lifeboat. In the autumn of 1899, his first demonstration in the United States took place. Marconi had sailed to the U.S. at the invitation of The New York Herald newspaper to cover the America's Cup international yacht races off Sandy Hook, New Jersey. The transmission was done aboard the SS Ponce, a passenger ship of the Porto Rico Line. Marconi left for England on 8 November 1899 on the American Line's , and he and his assistants installed wireless equipment aboard during the voyage. Prior to this voyage the Second Boer War had begun, and Marconi's wireless would bring news of the conflict to passengers at the request of "some of the officials of the American line." On 15 November the SS Saint Paul became the first ocean liner to report her imminent return to Great Britain by wireless when Marconi's Royal Needles Hotel radio station contacted her 66 nautical miles off the English coast. The first Transatlantic Times, a newspaper containing wireless transmission news from the Needles Station at the Isle of Wight, was published on board the SS Saint Paul prior to its arrival. Transatlantic transmissions At the turn of the 20th century, Marconi began investigating a means to signal across the Atlantic to compete with the transatlantic telegraph cables. Marconi established a wireless transmitting station at Marconi House, Rosslare Strand, County Wexford, in 1901 to act as a link between Poldhu in Cornwall, England, and Clifden in Connemara, County Galway, Ireland. He soon made the announcement that the message was received at Signal Hill in St. John's, Newfoundland (now part of Canada), on 12 December 1901, using a kite-supported antenna for reception—signals transmitted by the company's new high-power station at Poldhu, Cornwall. The distance between the two points was about . It was heralded as a great scientific advance, yet there also was—and continues to be—considerable scepticism about this claim. The exact wavelength used is not known, but it is fairly reliably determined to have been in the neighbourhood of 350 meters (frequency ≈ 850 kHz). The tests took place at a time of day during which the entire transatlantic path was in daylight. It is now known (although Marconi did not know then) that this was the worst possible choice. At this medium wavelength, long-distance transmission in the daytime is not possible because of heavy absorption of the skywave in the ionosphere. It was not a blind test; Marconi knew in advance to listen for a repetitive signal of three clicks, signifying the Morse code letter S. The clicks were reported to have been heard faintly and sporadically. There was no independent confirmation of the reported reception, and the transmissions were difficult to distinguish from atmospheric noise. A detailed technical review of Marconi's early transatlantic work appears in John S. Belrose's work of 1995. The Poldhu transmitter was a two-stage circuit.Belrose, John S. (5 September 1995) "Fessenden and Marconi: Their Differing Technologies and Transatlantic Experiments During the First Decade of this Century ". International Conference on 100 Years of Radio. Feeling challenged by sceptics, Marconi prepared a better organised and documented test. In February 1902, the SS Philadelphia sailed west from Great Britain with Marconi aboard, carefully recording signals sent daily from the Poldhu station. The test results produced coherer-tape reception up to , and audio reception up to . The maximum distances were achieved at night, and these tests were the first to show that radio signals for medium wave and longwave transmissions travel much farther at night than in the day. During the daytime, signals had been received up to only about , less than half of the distance claimed earlier at Newfoundland, where the transmissions had also taken place during the day. Because of this, Marconi had not fully confirmed the Newfoundland claims, although he did prove that radio signals could be sent for hundreds of kilometres (miles), despite some scientists' belief that they were limited essentially to line-of-sight distances. On 17 December 1902, a transmission from the Marconi station in Glace Bay, Nova Scotia, Canada, became the world's first radio message to cross the Atlantic from North America. In 1901, Marconi built a station near South Wellfleet, Massachusetts, that sent a message of greetings on 18 January 1903 from United States President Theodore Roosevelt to King Edward VII of the United Kingdom. However, consistent transatlantic signalling was difficult to establish. Marconi began to build high-powered stations on both sides of the Atlantic to communicate with ships at sea, in competition with other inventors. In 1904, he established a commercial service to transmit nightly news summaries to subscribing ships, which could incorporate them into their on-board newspapers. A regular transatlantic radio-telegraph service was finally begun on 17 October 1907Second Test of the Marconi Over-Ocean Wireless System Proved Entirely Successful . Sydney Daily Post. 24 October 1907. between Clifden, Ireland, and Glace Bay, but even after this the company struggled for many years to provide reliable communication to others. Titanic The role played by Marconi Co. wireless in maritime rescues raised public awareness of the value of radio and brought fame to Marconi, particularly the sinking of RMS Titanic on 15 April 1912 and RMS Lusitania on 7 May 1915. RMS Titanic radio operators Jack Phillips and Harold Bride were not employed by the White Star Line but by the Marconi International Marine Communication Company. After the sinking of the ocean liner, survivors were rescued by the RMS Carpathia of the Cunard Line. Carpathia took a total of 17 minutes to both receive and decode the SOS signal sent by Titanic. There was a distance of 58 miles between the two ships. When Carpathia docked in New York, Marconi went aboard with a reporter from The New York Times to talk with Bride, the surviving operator. After this incident, Marconi gained popularity and became more recognised for his contributions to the field of radio and wireless technology. On 18 June 1912, Marconi gave evidence to the Court of Inquiry into the loss of Titanic regarding the marine telegraphy's functions and the procedures for emergencies at sea. Britain's Postmaster-General summed up, referring to the Titanic disaster: "Those who have been saved, have been saved through one man, Mr. Marconi ... and his marvellous invention." Marconi was offered free passage on Titanic before she sank, but had taken Lusitania three days earlier. As his daughter Degna later explained, he had paperwork to do and preferred the public stenographer aboard that vessel. Continuing work Over the years, the Marconi companies gained a reputation for being technically conservative, in particular by continuing to use inefficient spark-transmitter technology, which could be used only for radio-telegraph operations, long after it was apparent that the future of radio communication lay with continuous-wave transmissions which were more efficient and could be used for audio transmissions. Somewhat belatedly, the company did begin significant work with continuous-wave equipment beginning in 1915, after the introduction of the oscillating vacuum tube (valve). The New Street Works factory in Chelmsford was the location for the first entertainment radio broadcasts in the United Kingdom in 1920, employing a vacuum tube transmitter and featuring Dame Nellie Melba. In 1922, regular entertainment broadcasts commenced from the Marconi Research Centre at Great Baddow, forming the prelude to the BBC, and he spoke of the close association of aviation and wireless telephony in that same year at a private gathering with Florence Tyzack Parbury, and even spoke of interplanetary wireless communication. In 1924, the Marconi Company co-established the Unione Radiofonica Italiana (now RAI). Later years In 1914, Marconi was made a Senator in the Senate of the Kingdom of Italy and appointed Honorary Knight Grand Cross of the Royal Victorian Order in the UK. During World War I, Italy joined the Allied side of the conflict, and Marconi was placed in charge of the Italian military's radio service. He attained the rank of lieutenant in the Italian Royal Army and of commander in the Regia Marina. In 1929, he was made a marquess by King Victor Emmanuel III. While helping to develop microwave technology, the Marchese Marconi suffered nine heart attacks in the span of three years preceding his death. Marconi died in Rome on 20 July 1937 at age 63, following the ninth, fatal, heart attack, and Italy held a state funeral for him. As a tribute, shops on the street where he lived were "Closed for national mourning". In addition, at 6 pm the next day, the time designated for the funeral, transmitters around the world observed two minutes of silence in his honour. The British Post Office also sent a message requesting that all broadcasting ships honour Marconi with two minutes of broadcasting silence. His remains are housed in the Mausoleum of Guglielmo Marconi in the grounds of Villa Griffone at Sasso Marconi, Emilia-Romagna, which assumed that name in his honour in 1938. In 1943, Marconi's elegant sailing yacht, the Elettra, was commandeered and refitted as a warship by the German Navy. She was sunk by the RAF on 22 January 1944. After the war, the Italian Government tried to retrieve the wreckage, to rebuild the boat, and the wreckage was removed to Italy. Eventually, the idea was abandoned, and the wreckage was cut into pieces which were distributed amongst Italian museums. In 1943, the Supreme Court of the United States handed down a decision on Marconi's radio patents restoring some of the prior patents of Oliver Lodge, John Stone Stone, and Nikola Tesla. The decision was not about Marconi's original radio patents and the court declared that their decision had no bearing on Marconi's claim as the first to achieve radio transmission, just that since Marconi's claim to certain patents was questionable, he could not claim infringement on those same patents. There are claims the high court was trying to nullify a World War I claim against the United States government by the Marconi Company via simply restoring the non-Marconi prior patent. Personal life Marconi was a friend of Charles van Raalte and his wife Florence, the owners of Brownsea Island; and of Margherita, their daughter, and in 1904 he met her Irish friend, The Hon. Beatrice O'Brien (1882–1976), a daughter of The 14th Baron Inchiquin. On 16 March 1905, Beatrice O'Brien and Marconi were married, and spent their honeymoon on Brownsea Island. They had three daughters, Degna (1908–1998), Gioia (1916–1996), and Lucia (born and died 1906), and a son, Giulio, 2nd Marchese Marconi (1910–1971). In 1913, the Marconi family returned to Italy and became part of Rome society. Beatrice served as a lady-in-waiting to Queen Elena. At Marconi's request, his marriage to Beatrice was annulled on 27 April 1927, so he could remarry. Marconi and Beatrice had divorced on 12 February 1924 in the free city of Fiume (Rijeka). On 12 June 1927 Marconi went on to marry Maria Cristina Bezzi-Scali (2 April 1900 – 15 July 1994), the only daughter of Francesco, Count Bezzi-Scali. To do this he had to be confirmed in the Catholic faith and became a devout member of the Church. He was baptised Catholic but had been brought up as a member of the Anglican Church. On 12 June 1927, Marconi married Maria Cristina in a civil service, with a religious ceremony performed on 15 June. Marconi was 53 years old and Maria Cristina was 26. They had one daughter, Maria Elettra Elena Anna (born 1930), who married Prince Carlo Giovannelli (1942–2016) in 1966; they later divorced. For unexplained reasons, Marconi left his entire fortune to his second wife and their only child, and nothing to the children of his first marriage. Marconi wanted to personally introduce in 1931 the first radio broadcast of a Pope, Pius XI, and did announce at the microphone: "With the help of God, who places so many mysterious forces of nature at man's disposal, I have been able to prepare this instrument which will give to the faithful of the entire world the joy of listening to the voice of the Holy Father". Fascism Marconi joined the National Fascist Party in 1923. In 1930, Italian dictator Benito Mussolini appointed him President of the Royal Academy of Italy, which made Marconi a member of the Fascist Grand Council. Marconi was an apologist for fascist ideology and actions such as the Italian invasion of Ethiopia in the Second Italo-Abyssinian War. In his lecture he stated: "I reclaim the honour of being the first fascist in the field of radiotelegraphy, the first who acknowledged the utility of joining the electric rays in a bundle, as Mussolini was the first in the political field who acknowledged the necessity of merging all the healthy energies of the country into a bundle, for the greater greatness of Italy". In 2002 researcher Annalisa Capristo found documents in the archives of Rome which showed that during his time as the President of the Royal Academy of Italy, Marconi had marked by hand Jewish applicants' records with an "E", where in the Italian language word for Jew is "Ebreo". Not one Jew was allowed to join during Marconi's tenure as president from 1930, three years before Adolf Hitler took power in Germany and eight years before Benito Mussolini's race laws brought his regime's antisemitism into the open. Following publication of Capristo's article "The Exclusion of Jews From the Academy of Italy" published in the Israel Monthly Review, historians were divided over whether the discrimination was the personal initiative of a scientist who considered Jews inferior or whether it was the action of a man too weak to oppose the regime's edicts. Legacy and honours Archives A large collection of Marconi artefacts was held by The General Electric Company, plc (GEC) of the United Kingdom which later renamed itself Marconi plc and Marconi Corporation plc. In December 2004 the extensive Marconi Collection, held at the former Marconi Research Centre at Great Baddow, Chelmsford, Essex UK was donated to the nation by the Company via the University of Oxford. This consisted of the BAFTA award-winning MarconiCalling website, some 250+ physical artefacts and the massive ephemera collection of papers, books, patents and many other items. The artefacts are now held by The Museum of the History of Science and the ephemera Archives by the nearby Bodleian Library. Following three years' work at the Bodleian, an Online Catalogue to the Marconi Archives was released in November 2008. Orders and decorations Italian Knight of the Order of Merit for Labour (26 October 1902) Knight of the Civil Order of Savoy (1 June 1905) Grand Cordon of the Order of the Crown of Italy (7 April 1913; Grand Officer: 30 October 1902; Officer: 6 January 1898) Grand Cordon of the Order of Saints Maurice and Lazarus (14 January 1932; Grand Officer: 30 May 1912; Commander: 12 January 1902) Marquis of Marconi (17 July 1929) Others Grand Cross of the Order of Saint Anna of the Russia Empire (1902) Honorary Knight Grand Cross of the Royal Victorian Order of the United Kingdom (GCVO, 1914) Grand Cross of the Civil Order of Alfonso XII of Spain Grand Cordon of the Order of the Rising Sun of Japan (1933) Honours and awards In 1901, he was elected as a member of the American Philosophical Society. In 1903, Marconi also received the freedom of the City of Rome. In 1909, Marconi shared the Nobel Prize in Physics with Karl Ferdinand Braun for their "contributions to the development of wireless telegraphy" (radio communications). In 1914, Marconi was named senator by the king of Italy Vittorio Emanuele III In 1918, he was awarded the Franklin Institute's Franklin Medal. In 1920, he was awarded the IRE Medal of Honor, now the IEEE Medal of Honor. In 1931, he was awarded the John Scott Medal by the Franklin Institute and the City Council of Philadelphia. In 1934, he was awarded the Wilhelm Exner Medal. In 1974, Italy marked the birth centennial of Marconi with a circulating commemorative Lire 100 coin. In 1975, Marconi was inducted into the National Inventors Hall of Fame. In 1978, Marconi was inducted into the NAB Broadcasting Hall of Fame. In 1988, the Radio Hall of Fame (Museum of Broadcast Communications, Chicago) inducted Marconi as a Pioneer (soon after the inception of its awards). In 1990, the Bank of Italy issued a Lire 2,000 banknote featuring his portrait on the front and on the back his accomplishments. In 2001, Great Britain released a commemorative £2 coin celebrating the 100th anniversary of Marconi's first wireless communication. Marconi's early experiments in wireless telegraphy were the subject of two IEEE Milestones; one in Switzerland in 2003 and most recently in Italy in 2011. In 2009, Italy issued a commemorative silver 10 Euro coin honouring the centennial of Marconi's Nobel Prize. In 2009, he was inducted into the New Jersey Hall of Fame. The Dutch radio academy bestows the annually for outstanding radio programmes, presenters and stations. The National Association of Broadcasters (US) bestows the annual NAB Marconi Radio Awards also for outstanding radio programmes and stations. Tributes A funerary monument to the effigy of Marconi can be seen in the Basilica of Santa Croce, Florence, but his remains are in the Mausoleum of Guglielmo Marconi in Sasso Marconi, Italy. His former villa, adjacent to the mausoleum is the Marconi Museum (Italy) with much of his equipment. A Guglielmo Marconi sculpture by Attilio Piccirilli stands in Washington, D.C. A granite obelisk stands on the cliff top near the site of Marconi's Marconi's Poldhu Wireless Station in Cornwall, commemorating the first transatlantic transmission. Marconi Plaza Park, an urban park square named after the inventor in 1937, is located Philadelphia, Pennsylvania at Oregon Ave and South Broad Street. It includes a later 1975 bronze statue of Marconi erected on the east side. Places and organisations named after Marconi Outer space The asteroid 1332 Marconia is named in his honour. A large crater on the far side of the moon is also named after him. EuropeItaly Bologna Guglielmo Marconi Airport (IATA: BLQ – ICAO: LIPE), of Bologna, is named after Marconi, its native son. Ponte Guglielmo Marconi, bridge that connects Piazza Augusto Righi with Piazza Tommaso Edison, in Rome OceaniaAustralia Australian football (soccer) and social club Marconi Stallions. North America Canada''' The Marconi's Wireless Telegraph Company of Canada (now CMC Electronics and Ultra Electronics), of Montreal, Quebec, Canada, was created in 1903 by Guglielmo Marconi. In 1925 the company was renamed to the 'Canadian Marconi Company', which was acquired by English Electric in 1953. The company name changed again to CMC Electronics Inc. (French: CMC Électronique) in 2001. In 2002, the company historical radio business was sold to Ultra Electronics to become Ultra Electronics TCS Inc., now doing business as Ultra Communications. Both CMC Electronics and Ultra Communications are still located in Montreal. The Marconi National Historic Sites of Canada was created by Parks Canada as a tribute to Marconi's vision in the development of radio telecommunications. The first official wireless message was sent from this location by the Atlantic Ocean to England in 1902. The museum site is located in Glace Bay, Nova Scotia, at Table Head on Timmerman Street. United States California Marconi Conference Center and State Historic Park, site of the transoceanic Marshall Receiving Station, Marshall. Hawaii Marconi Wireless Telegraphy Station on Oahu's North Shore, briefly the world's most powerful telegraph station. Massachusetts Marconi Beach in Wellfleet, Massachusetts, part of the Cape Cod National Seashore, located near the site of his first transatlantic wireless signal from the United States to Britain. There are still remnants of the wireless tower at this beach and at Forest Road Beach in Chatham, Massachusetts. New Jersey New Brunswick Marconi Station, now the Guglielmo Marconi Memorial Plaza in Somerset, NJ. President Woodrow Wilson's Fourteen Points speech was transmitted from the site in 1918. Belmar Marconi Station, now the InfoAge Science History Center in Wall Township, NJ. The Marconi Wireless Company of America, the world's first radio company, was incorporated in Roselle Park New Jersey, on West Westfield Avenue, on November 22, 1899. New York La Scuola d'Italia Guglielmo Marconi on New York City's Upper East Side. Pennsylvania Marconi Plaza, Philadelphia, Pennsylvania. Roman terrace-styled plaza originally designed by the architects Olmsted Brothers in 1914–1916, built as the grand entrance for the 1926 Sesquicentennial Exposition and renamed to honour Marconi. Patents British patents British patent No. 12,039 (1897) "Improvements in Transmitting Electrical impulses and Signals, and in Apparatus therefor". Date of Application 2 June 1896; Complete Specification Left, 2 March 1897; Accepted, 2 July 1897 (later claimed by Oliver Lodge to contain his own ideas which he failed to patent). British patent No. 7,777 (1900) "Improvements in Apparatus for Wireless Telegraphy". Date of Application 26 April 1900; Complete Specification Left, 25 February 1901; Accepted, 13 April 1901. British patent No. 10245 (1902) British patent No. 5113 (1904) "Improvements in Transmitters suitable for Wireless Telegraphy". Date of Application 1 March 1904; Complete Specification Left, 30 November 1904; Accepted, 19 January August 1905. British patent No. 21640 (1904) "Improvements in Apparatus for Wireless Telegraphy". Date of Application 8 October 1904; Complete Specification Left, 6 July 1905; Accepted, 10 August 1905. British patent No. 14788 (1904) "Improvements in or relating to Wireless Telegraphy". Date of Application 18 July 1905; Complete Specification Left, 23 January 1906; Accepted, 10 May 1906. US patents "Transmitting electrical signals", (using Ruhmkorff coil and Morse code key) filed December 1896, patented July 1897 "Apparatus employed in wireless telegraphy". "Apparatus employed in wireless telegraphy". "Apparatus employed in wireless telegraphy". "Apparatus employed in wireless telegraphy". "Apparatus employed in wireless telegraphy". "Apparatus employed in wireless telegraphy". "Apparatus employed in wireless telegraphy". "Receiver for electrical oscillations". "Apparatus for wireless telegraphy" (later practical version of system) "Wireless telegraphy system". Filed 19 November 1901; Issued 19 April 1904. "Wireless signaling system". Filed 10 September 1903; Issued 24 May 1904. "Apparatus for wireless telegraphy" (Four tuned system; this innovation was predated by N. Tesla, O. Lodge, and J. S. Stone) "Wireless telegraphy". Filed 13 October 1903 "Wireless telegraphy". Filed 13 October 1903; Issued 13 June 1905. "Wireless telegraphy". Filed 28 November 1902; Issued 14 April 1908. "Wireless telegraphy". "Detecting electrical oscillations". Filed 2 February 1903; Issued 14 April 1908. "Wireless telegraphy". Filed 2 February 1903; Issued 14 April 1908. "Wireless signaling system". Filed 9 August 1906; Issued 8 June 1909. "Transmitting apparatus for wireless telegraphy". Filed 10 April 1908; Issued 28 September 1909. "Apparatus for wireless telegraphy". "Apparatus for wireless telegraphy". Filed 10 April 1908; Issued 28 September 1909. "Apparatus for wireless telegraphy". Filed 31 March 1909; Issued 12 April 1910. "Transmitting apparatus for wireless telegraphy". Filed 15 July 1910; Issued 11 July 1911. "Means for generating alternating electric currents". Filed 27 January 1914; Issued 7 July 1914. "Transmitting apparatus for use in wireless telegraphy and telephony". Filed 31 December 1913; Issued 15 May 1917. "Wireless telegraph transmitter". "Electric accumulator". Filed 9 March 1918 "Transmitter for wireless telegraphy". Filed 20 July 1908; Issued 3 August 1915. "Thermionic valve". Filed 14 October 1926; Issued 20 November 1934. Reissued (US) "Transmitting electrical impulses and signals and in apparatus, there-for". Filed 1 April 1901; Issued 4 June 1901. See also History of radio Jagadish Chandra Bose List of people on stamps of Ireland List of covers of Time magazine during the 1920s – 6 December 1926 Marconi's law References Sources Further reading Relatives and company publications Bussey, Gordon, Marconi's Atlantic Leap, Marconi Communications, 2000. Isted, G.A., Guglielmo Marconi and the History of Radio – Part I, General Electric Company, p.l.c., GEC Review, Volume 7, No. 1, p45, 1991, Isted, G.A., Guglielmo Marconi and the History of Radio – Part II, General Electric Company, p.l.c., GEC Review, Volume 7, No. 2, p110, 1991, Marconi, Degna, My Father, Marconi, James Lorimer & Co, 1982. (Italian version): Marconi, mio padre, Di Renzo Editore, 2008, Marconi's Wireless Telegraph Company, Year book of wireless telegraphy and telephony, London: Published for the Marconi Press Agency Ltd., by the St. Catherine Press / Wireless Press. LCCN 14017875 sn 86035439 Simons, R.W., Guglielmo Marconi and Early Systems of Wireless Communication, General Electric Company, p.l.c., GEC Review, Volume 11, No. 1, p37, 1996, Scholarly studies Ahern, Steve (ed), Making Radio (2nd Edition) Allen & Unwin, Sydney, 2006 . Aitken, Hugh G. J., Syntony and Spark: The Origins of Radio, New York: John Wiley & Sons, 1976. Aitken, Hugh G. J., The Continuous Wave: Technology and American Radio, 1900–1932, Princeton, New Jersey: Princeton University Press, 1985. . Anderson, Leland I., Priority in the Invention of Radio – Tesla vs. Marconi Baker, W. J., A History of the Marconi Company, 1970. Brodsky, Ira. The History of Wireless: How Creative Minds Produced Technology for the Masses (Telescope Books, 2008) Cheney, Margaret, Tesla: Man Out of Time Laurel Publishing, 1981. Chapter 7, esp pp 69, re: published lectures of Tesla in 1893, copied by Marconi. Clark, Paddy, "Marconi's Irish Connections Recalled," published in 100 Years of Radio, IEE Conference Publication 411, 1995. Coe, Douglas and Kreigh Collins (ills), Marconi, pioneer of radio, New York, J. Messner, Inc., 1943. LCCN 43010048 Garratt, G. R. M., The early history of radio: from Faraday to Marconi, London, Institution of Electrical Engineers in association with the Science Museum, History of technology series, 1994. LCCN gb 94011611 Geddes, Keith, Guglielmo Marconi, 1874–1937, London : H.M.S.O., A Science Museum booklet, 1974. LCCN 75329825 (ed. Obtainable in the United States. from Pendragon House Inc., Palo Alto, California.) Hancock, Harry Edgar, Wireless at sea; the first fifty years: A history of the progress and development of marine wireless communications written to commemorate the jubilee of the Marconi International Marine Communication Company, Limited, Chelmsford, Eng., Marconi International Marine Communication Co., 1950. LCCN 51040529 /L Homer, Peter and O'Connor, Finbar, Marconi Wireless Radio Station: Malin Head from 1902, 2014. Hughes, Michael and Bosworth, Katherine, Titanic Calling : Wireless Communications During the Great Disaster, Oxford, WorldCat.org, 2012, Janniello, Maria Grace, Monteleone, Franco and Paoloni, Giovanni (eds) (1996), One hundred years of radio: From Marconi to the future of the telecommunications. Catalogue of the extension, Venice: Marsilio. Jolly, W. P., Marconi, 1972. Larson, Erik, Thunderstruck, New York: Crown Publishers, 2006. A comparison of the lives of Hawley Harvey Crippen and Marconi. Crippen was a murderer whose Transatlantic escape was foiled by the new invention of shipboard radio. MacLeod, Mary K., Marconi: The Canada Years – 1902–1946, Halifax, Nova Scotia: Nimbus Publishing Limited, 1992, Masini, Giancarlo, Guglielmo Marconi, Turin: Turinese typographical-publishing union, 1975. LCCN 77472455 (ed. Contains 32 tables outside of the text) Mason, H. B. (1908). Encyclopaedia of ships and shipping, Wireless Telegraphy. London: Shipping Encyclopaedia. 1908. Paul M. Hawkins - "Point to Point - A History of International Telecommunications During the Radio Years" ISBN 978-178719-6278 pub. by New Generation Publishing. Paul M. Hawkins & Paul G. Reyland - "Marconi's Wireless Telegraph Stations in Essex - The Centenary of Brentwood and Ongar Radio Stations" ISBN 978-180369-3828 by - pub.2022 by New Generation Publishing. Raboy, Marc. Marconi: The Man Who Networked the World (Oxford University Press, 2016) 872 pp. online review Stone, Ellery W., Elements of Radiotelegraphy Weightman, Gavin, Signor Marconi's magic box: the most remarkable invention of the 19th century & the amateur inventor whose genius sparked a revolution, 1st Da Capo Press ed., Cambridge, MA : Da Capo Press, 2003. Winkler, Jonathan Reed. Nexus: Strategic Communications and American Security in World War I. (Cambridge, MA: Harvard University Press, 2008). Account of rivalry between Marconi's firm and the United States government during World War I. External links Wikimedia General achievements including the Nobel Lecture, 11 December 1909 Wireless Telegraphic Communication Marconi il 5 marzo 1896, presenta a Londra la prima richiesta provvisoria di brevetto, col numero 5028 e col titolo "Miglioramenti nella telegrafia e relativi apparati" (Great Britain and France between 1896 and 1924) List of British and French patents (1896–1924) The first patent application number 5028 of 5 March 1896 (Provisional deprivation) Foundations and academics University of Oxford Introduction to the Online Catalogue of the Marconi Collection University of Oxford Online Catalogue of the Marconi Archives Guglielmo Marconi Foundation, Pontecchio Marconi, Bologna, Italy Galileo Legacy Foundation: pictures of the Dedication of the Guglielmo Marconi Square, Johnston RI United States Dedication Photos History of Marconi House, Marconi House, Strand / Aldwych, London. Multimedia and books MarconiCalling – The Life, Science and Achievements of Guglielmo Marconi, part of the Marconi Collection at the University of Oxford Canadian Heritage Minute featuring Marconi Guglielmo Marconi documentary , narrated by Walter Cronkite Review of Signor Marconi's Magic Box Transatlantic "signals" and radio Robert (Bob) White, Guglielmo Marconi – Aerial Assistance with a Kite. Bridging the Atlantic By Wireless Signal – 12 December 1901. Kiting, The Journal of the American Kitefliers Association. Vol. 23, Issue 5 – Winter 2002. November 2001 Faking the Waves, 1901 Marconi and "wireless telegraphy" using kites Keys and "signals" Sparks Telegraph Key Review An exhaustive listing of wireless telegraph key manufacturers including photos of most Marconi keys United States Senate Inquiry into the Titanic disaster – Testimony of Guglielmo Marconi Priority of invention vs Tesla PBS: Marconi and Tesla: Who invented radio? United States Supreme Court, Marconi Wireless Telegraph co. of America v. United States. 320 U.S. 1. Nos. 369, 373. Argued 9–12 April 1943. Decided 21 June 1943. 21st Century Books: Priority in the Invention of Radio – Tesla vs. Marconi Personal Information about Marconi and his yacht Elettra I diari di laboratorio di Guglielmo Marconi (The diaries of laboratory Guglielmo Marconi.) Comitato Guglielmo Marconi International, Bologna, Italy (Marconi's voice) August 1914 photo article on Marconi Belmar station in Wall, NJ, InfoAge. (See also, Marconi Period of Significance Historic Buildings.) Marconi, Guglielmo: Statue north of Meridian Hill Park in Washington , D.C. by Attilio Piccirilli Other Guglielmo Marconi, 2000 Italian Lire (1990) 1874 births 1937 deaths Engineers from Bologna Italian people of Scottish descent Italian people of Irish descent Nobility from Bologna Italian Roman Catholics Converts to Roman Catholicism from Anglicanism Members of the Grand Council of Fascism Members of the Senate of the Kingdom of Italy 20th-century Italian physicists European amateur radio operators Experimental physicists Italian electrical engineers Italian emigrants to the United States Italian expatriates in England Italian fascists 19th-century Italian inventors Radio pioneers Members of the Pontifical Academy of Sciences Members of the Royal Academy of Italy Foreign associates of the National Academy of Sciences Rectors of the University of St Andrews Italian military personnel of World War I Nobel laureates in Physics Italian Nobel laureates John Fritz Medal recipients Recipients of the Order of the Crown (Italy) Recipients of the Order of Merit for Labour Recipients of the Order of St. Anna Honorary Knights Grand Cross of the Royal Victorian Order IEEE Medal of Honor recipients Amateur radio people Businesspeople from Bologna Recipients of the Matteucci Medal History of radio
12236
https://en.wikipedia.org/wiki/GDP%20%28disambiguation%29
GDP (disambiguation)
GDP, short for gross domestic product, is the basic measure of a country's overall economic output. GDP may also refer to: GDP (chemotherapy), a chemotherapy treatment regimen GDP (musician), an American hip hop musician from New Jersey Giant depolarizing potentials, the first type of electrical activity of developing brain Gibraltar Defence Police, a civil police force which guards and enforces law on Ministry of Defence installations in Gibraltar Gidea Park railway station, National Rail station code GDP Good distribution practice, the guidelines for the proper distribution of medicinal products for human use Global Drifter Program, a program that was aiming to collect measurements of surface ocean currents, sea surface temperature and sea-level atmospheric pressure using drifters. Good documentation practice, pharmaceutical description of standards by which documents are created and maintained Ground delay program, a traffic flow initiative for aviation in the United States Guanosine diphosphate, a nucleotide Grand Ducal Police, the national police force of Luxembourg GdP may refer to: (), a trade union of police employees in Germany
12422
https://en.wikipedia.org/wiki/List%20of%20gymnasts
List of gymnasts
Gymnasts are people who participate in the sport of gymnastics. This sport contains disciplines that include, but are not limited to: {{columns-list|colwidth=22em| acrobatic gymnastics aerobic gymnastics<ref name="FIGAerobics">{{cite web|url=http://www.fig-gymnastics.com/vsite/vnavsite/page/directory/0,10853,5187-188475-205697-nav-list,00.html|title=Fédération Internationale de Gymnastique : About Aerobic|publisher=Fédération Internationale de Gymnastique|access-date=18 November 2012|quote=Aerobic has not only been a great form of physical training for the general public since the end of the 1980's, but also a top-level competitive sport. Aerobic Gymnastics presents dynamic moves, strength, flexibility, co-ordination and musicality in a routine, lasting up to 1 minute 50 seconds (1’30 for all categories from 2013).|archive-url=https://web.archive.org/web/20090323030420/http://www.fig-gymnastics.com/vsite/vnavsite/page/directory/0,10853,5187-188475-205697-nav-list,00.html|archive-date=2009-03-23|url-status=dead}}</ref> artistic gymnastics rhythmic gymnastics trampoline gymnastics tumbling}} This list is of those who are considered to be notable in their chosen discipline. See gymnasium (ancient Greece) for the origin of the word gymnast'' from gymnastikos. Artistic gymnasts List of women's artistic gymnasts List of men's artistic gymnasts Rhythmic gymnasts Female (rhythmic) Trampoline gymnasts Female (trampoline) Male (trampoline) See also International Gymnastics Hall of Fame List of Olympic medalists in gymnastics (men) List of Olympic medalists in gymnastics (women) List of current female artistic gymnasts Notes References External links World Ranking (1952-2005) Summary of Canadian Gymnastic Record University of Michigan gymnasts Gymnastics-related lists
12562
https://en.wikipedia.org/wiki/Gregor%20Mendel
Gregor Mendel
Gregor Johann Mendel OSA (; ; 20 July 1822 – 6 January 1884) was an Austrian biologist, meteorologist, mathematician, Augustinian friar and abbot of St. Thomas' Abbey in Brünn (Brno), Margraviate of Moravia. Mendel was born in a German-speaking family in the Silesian part of the Austrian Empire (today's Czech Republic) and gained posthumous recognition as the founder of the modern science of genetics. Though farmers had known for millennia that crossbreeding of animals and plants could favor certain desirable traits, Mendel's pea plant experiments conducted between 1856 and 1863 established many of the rules of heredity, now referred to as the laws of Mendelian inheritance. Mendel worked with seven characteristics of pea plants: plant height, pod shape and color, seed shape and color, and flower position and color. Taking seed color as an example, Mendel showed that when a true-breeding yellow pea and a true-breeding green pea were cross-bred their offspring always produced yellow seeds. However, in the next generation, the green peas reappeared at a ratio of 1 green to 3 yellow. To explain this phenomenon, Mendel coined the terms "recessive" and "dominant" in reference to certain traits. In the preceding example, the green trait, which seems to have vanished in the first filial generation, is recessive and the yellow is dominant. He published his work in 1866, demonstrating the actions of invisible "factors"—now called genes—in predictably determining the traits of an organism. The profound significance of Mendel's work was not recognized until the turn of the 20th century (more than three decades later) with the rediscovery of his laws. Erich von Tschermak, Hugo de Vries and Carl Correns independently verified several of Mendel's experimental findings in 1900, ushering in the modern age of genetics. Early life and education Mendel was born into a German-speaking family in Heinzendorf bei Odrau (now Hynčice, Czech Republic), at the Moravian-Silesian border, Austrian Empire. He was the son of Anton and Rosine (Schwirtlich) Mendel and had one older sister, Veronika, and one younger, Theresia. They lived and worked on a farm which had been owned by the Mendel family for at least 130 years (the house where Mendel was born is now a museum devoted to Mendel). During his childhood, Mendel worked as a gardener and studied beekeeping. As a young man, he attended gymnasium in Troppau (now Opava, Czech Republic). He had to take four months off during his gymnasium studies due to illness. From 1840 to 1843, he studied practical and theoretical philosophy and physics at the Philosophical Institute of the University of Olmütz (now Olomouc, Czech Republic), taking another year off because of illness. He also struggled financially to pay for his studies, and Theresia gave him her dowry. Later he helped support her three sons, two of whom became doctors. He became a monk in part because it enabled him to obtain an education without having to pay for it himself. As the son of a struggling farmer, the monastic life, in his words, spared him the "perpetual anxiety about a means of livelihood." Born Johann Mendel, he was given the name Gregor ( in Czech) when he joined the Order of Saint Augustine. Academic career When Mendel entered the Faculty of Philosophy, the Department of Natural History and Agriculture was headed by Johann Karl Nestler who conducted extensive research of hereditary traits of plants and animals, especially sheep. Upon recommendation of his physics teacher Friedrich Franz, Mendel entered the Augustinian St Thomas's Abbey in Brünn (now Brno, Czech Republic) and began his training as a priest. Mendel worked as a substitute high school teacher. In 1850, he failed the oral part, the last of three parts, of his exams to become a certified high school teacher. In 1851, he was sent to the University of Vienna to study under the sponsorship of Abbot so that he could get more formal education. At Vienna, his professor of physics was Christian Doppler. Mendel returned to his abbey in 1853 as a teacher, principally of physics. In 1854 he met Aleksander Zawadzki who encouraged his research in Brno. In 1856, he took the exam to become a certified teacher and again failed the oral part. In 1867, he replaced Napp as abbot of the monastery. After he was elevated as abbot in 1868, his scientific work largely ended, as Mendel became overburdened with administrative responsibilities, especially a dispute with the civil government over its attempt to impose special taxes on religious institutions. Mendel died on 6 January 1884, at the age of 61, in Brünn, Moravia, Austria-Hungary (now Czech Republic), from chronic nephritis. Czech composer Leoš Janáček played the organ at his funeral. After his death, the succeeding abbot burned all papers in Mendel's collection, to mark an end to the disputes over taxation. The exhumation of Mendel's corpse in 2021 delivered some physiognomic details like body height (). His genome was analysed, revealing that Mendel also suffered from heart problems. Contributions Experiments on plant hybridization Mendel, known as the "father of modern genetics", chose to study variation in plants in his monastery's experimental garden. Mendel was assisted in his experimental design by Aleksander Zawadzki while his superior abbot Napp wrote to discourage him, saying that the Bishop giggled when informed of the detailed genealogies of peas. After initial experiments with pea plants, Mendel settled on studying seven traits that seemed to be inherited independently of other traits: seed shape, flower color, seed coat tint, pod shape, unripe pod color, flower location, and plant height. He first focused on seed shape, which was either angular or round. Between 1856 and 1863 Mendel cultivated and tested some 28,000 plants, the majority of which were pea plants (Pisum sativum). This study showed that, when true-breeding different varieties were crossed to each other (e.g., tall plants fertilized by short plants), in the second generation, one in four pea plants had purebred recessive traits, two out of four were hybrids, and one out of four were purebred dominant. His experiments led him to make two generalizations, the Law of Segregation and the Law of Independent Assortment, which later came to be known as Mendel's Laws of Inheritance. Initial reception of Mendel's work Mendel presented his paper, ("Experiments on Plant Hybridization"), at two meetings of the Natural History Society of Brno in Moravia on 8 February and 8 March 1865. It generated a few favorable reports in local newspapers, but was ignored by the scientific community. When Mendel's paper was published in 1866 in , it was seen as essentially about hybridization rather than inheritance, had little impact, and was cited only about three times over the next thirty-five years. His paper was criticized at the time, but is now considered a seminal work. Notably, Charles Darwin was not aware of Mendel's paper, and it is envisaged that if he had been aware of it, genetics as it exists now might have taken hold much earlier. Mendel's scientific biography thus provides an example of the failure of obscure, highly original innovators to receive the attention they deserve. Rediscovery of Mendel's work About forty scientists listened to Mendel's two groundbreaking lectures, but it would appear that they failed to understand his work. Later, he also carried on a correspondence with Carl Nägeli, one of the leading biologists of the time, but Nägeli too failed to appreciate Mendel's discoveries. At times, Mendel must have entertained doubts about his work, but not always: "My time will come," he reportedly told a friend, Gustav von Niessl. During Mendel's lifetime, most biologists held the idea that all characteristics were passed to the next generation through blending inheritance, in which the traits from each parent are averaged. Instances of this phenomenon are now explained by the action of multiple genes with quantitative effects. Charles Darwin tried unsuccessfully to explain inheritance through a theory of pangenesis. It was not until the early 20th century that the importance of Mendel's ideas was realized. By 1900, research aimed at finding a successful theory of discontinuous inheritance rather than blending inheritance led to independent duplication of his work by Hugo de Vries and Carl Correns, and the rediscovery of Mendel's writings and laws. Both acknowledged Mendel's priority, and it is thought probable that de Vries did not understand the results he had found until after reading Mendel. Though Erich von Tschermak was originally also credited with rediscovery, this is no longer accepted because he did not understand Mendel's laws. Though de Vries later lost interest in Mendelism, other biologists started to establish modern genetics as a science. All three of these researchers, each from a different country, published their rediscovery of Mendel's work within a two-month span in the spring of 1900. Mendel's results were quickly replicated, and genetic linkage quickly worked out. Biologists flocked to the theory; even though it was not yet applicable to many phenomena, it sought to give a genotypic understanding of heredity which they felt was lacking in previous studies of heredity, which had focused on phenotypic approaches. Most prominent of these previous approaches was the biometric school of Karl Pearson and W. F. R. Weldon, which was based heavily on statistical studies of phenotype variation. The strongest opposition to this school came from William Bateson, who perhaps did the most in the early days of publicising the benefits of Mendel's theory (the word "genetics", and much of the discipline's other terminology, originated with Bateson). This debate between the biometricians and the Mendelians was extremely vigorous in the first two decades of the 20th century, with the biometricians claiming statistical and mathematical rigor, whereas the Mendelians claimed a better understanding of biology. Modern genetics shows that Mendelian heredity is in fact an inherently biological process, though not all genes of Mendel's experiments are yet understood. In the end, the two approaches were combined, especially by work conducted by R. A. Fisher as early as 1918. The combination, in the 1930s and 1940s, of Mendelian genetics with Darwin's theory of natural selection resulted in the modern synthesis of evolutionary biology. Other experiments Mendel began his studies on heredity using mice. He was at St. Thomas's Abbey but his bishop did not like one of his friars studying animal sex, so Mendel switched to plants. Mendel also bred bees in a bee house that was built for him, using bee hives that he designed. He also studied astronomy and meteorology, founding the 'Austrian Meteorological Society' in 1865. The majority of his published works were related to meteorology. Mendel also experimented with hawkweed (Hieracium) and honeybees. He published a report on his work with hawkweed, a group of plants of great interest to scientists at the time because of their diversity. However, the results of Mendel's inheritance study in hawkweeds was unlike his results for peas; the first generation was very variable and many of their offspring were identical to the maternal parent. In his correspondence with Carl Nägeli he discussed his results but was unable to explain them. It was not appreciated until the end of the nineteenth century that many hawkweed species were apomictic, producing most of their seeds through an asexual process. None of his results on bees survived, except for a passing mention in the reports of Moravian Apiculture Society. All that is known definitely is that he used Cyprian and Carniolan bees, which were particularly aggressive to the annoyance of other monks and visitors of the monastery such that he was asked to get rid of them. Mendel, on the other hand, was fond of his bees, and referred to them as "my dearest little animals". He also described novel plant species, and these are denoted with the botanical author abbreviation "Mendel". Mendelian paradox In 1936, Ronald Fisher, a prominent statistician and population geneticist, reconstructed Mendel's experiments, analyzed results from the F2 (second filial) generation and found the ratio of dominant to recessive phenotypes (e.g. yellow versus green peas; round versus wrinkled peas) to be implausibly and consistently too close to the expected ratio of 3 to 1. Fisher asserted that "the data of most, if not all, of the experiments have been falsified so as to agree closely with Mendel's expectations". Mendel's alleged observations, according to Fisher, were "abominable", "shocking", and "cooked". Other scholars agree with Fisher that Mendel's various observations come uncomfortably close to Mendel's expectations. A. W. F. Edwards, for instance, remarks: "One can applaud the lucky gambler; but when he is lucky again tomorrow, and the next day, and the following day, one is entitled to become a little suspicious". Three other lines of evidence likewise lend support to the assertion that Mendel's results are indeed too good to be true. Fisher's analysis gave rise to the Mendelian paradox: Mendel's reported data are, statistically speaking, too good to be true, yet "everything we know about Mendel suggests that he was unlikely to engage in either deliberate fraud or in unconscious adjustment of his observations". A number of writers have attempted to resolve this paradox. One attempted explanation invokes confirmation bias. Fisher accused Mendel's experiments as "biased strongly in the direction of agreement with expectation[...] to give the theory the benefit of doubt". In his 2004 article, J.W. Porteous concluded that Mendel's observations were indeed implausible. However, reproduction of the experiments has demonstrated that there is no real bias towards Mendel's data. Another attempt to resolve the Mendelian paradox notes that a conflict may sometimes arise between the moral imperative of a bias-free recounting of one's factual observations and the even more important imperative of advancing scientific knowledge. Mendel might have felt compelled "to simplify his data in order to meet real, or feared, editorial objections". Such an action could be justified on moral grounds (and hence provide a resolution to the Mendelian paradox), since the alternative—refusing to comply—might have retarded the growth of scientific knowledge. Similarly, like so many other obscure innovators of science, Mendel, a little known innovator of working-class background, had to "break through the cognitive paradigms and social prejudices" of his audience. If such a breakthrough "could be best achieved by deliberately omitting some observations from his report and adjusting others to make them more palatable to his audience, such actions could be justified on moral grounds". Daniel L. Hartl and Daniel J. Fairbanks reject outright Fisher's statistical argument, suggesting that Fisher incorrectly interpreted Mendel's experiments. They find it likely that Mendel scored more than 10 progeny, and that the results matched the expectation. They conclude: "Fisher's allegation of deliberate falsification can finally be put to rest, because on closer analysis it has proved to be unsupported by convincing evidence". In 2008 Hartl and Fairbanks (with Allan Franklin and AWF Edwards) wrote a comprehensive book in which they concluded that there were no reasons to assert Mendel fabricated his results, nor that Fisher deliberately tried to diminish Mendel's legacy. Reassessment of Fisher's statistical analysis, according to these authors, also disproves the notion of confirmation bias in Mendel's results. Commemoration Mount Mendel in New Zealand's Paparoa Range was named after him in 1970 by the Department of Scientific and Industrial Research. In celebration of his 200th birthday, Mendel's body was exhumed and his DNA sequenced. See also List of Roman Catholic cleric–scientists Mendel Museum of Genetics Mendel Polar Station in Antarctica Mendel University Brno Mendelian error The Gardener of God, an Italian docudrama about the life and works of Gregor Mendel References Further reading William Bateson On-line Facsimile Edition: Electronic Scholarly Publishing, Prepared by Robert Robbins Hugo Iltis, Gregor Johann Mendel. Leben, Werk und Wirkung. Berlin: J. Springer. 426 pages. (1924) Translated by Eden and Cedar Paul as Life of Mendel. New York: W. W. Norton & Co, 1932. 336 pages. New York: Hafner, 1966: London: George Allen & Unwin, 1966. Ann Arbor: University Microfilms International, 1976. Translated by Zhenyao Tan as Mên-tê-êrh chuan. Shanghai: Shang wu yin shu guan, 1924. 2 vols. in 1, 661 pp. Shanghai: Shang wu yin shu guan, Minguo 25 [1936]. Translated as Zasshu shokubutsu no kenkyū. Tsuketari Menderu shōden. Tōkyō: Iwanami Shoten, Shōwa 3 [1928]. 100 pp. Translated by Yuzuru Nagashima as Menderu no shōgai. Tōkyō: Sōgensha, Shōwa 17 [1942]. Menderu den. Tōkyō: Tōkyō Sōgensha, 1960. Robert Lock, Recent Progress in the Study of Variation, Heredity and Evolution, London, 1906 (1st Pub. 1905) Curt Stern and Sherwood ER (1966) The Origin of Genetics. refutes allegations about "data smoothing" James Walsh, Catholic Churchmen in Science, Philadelphia: Dolphin Press, 1906 External links 1913 Catholic Encyclopedia entry, "Mendel, Mendelism" Augustinian Abbey of St. Thomas at Brno Biography, bibliography and access to digital sources in the Virtual Laboratory of the Max Planck Institute for the History of Science Biography of Gregor Mendel GCSE student Gregor Mendel (1822–1884) Gregor Mendel Primary Sources Johann Gregor Mendel: Why his discoveries were ignored for 35 (72) years Masaryk University to rebuild Mendel’s greenhouse | Brno Now Mendel Museum of Genetics Mendel's Paper in English Online Mendelian Inheritance in Man A photographic tour of St. Thomas' Abbey, Brno, Czech Republic 1822 births 1884 deaths People from Nový Jičín District People from Austrian Silesia Silesian-German people Czech Roman Catholics Augustinian friars Austrian abbots Austrian biologists 19th-century Austrian botanists Austrian geneticists Botanists with author abbreviations Czech biologists Czech botanists Catholic clergy scientists Catholic philosophers 19th-century Austrian Roman Catholic priests Palacký University Olomouc alumni Deaths from nephritis Burials at Brno Central Cemetery
12710
https://en.wikipedia.org/wiki/Gary%20Snyder
Gary Snyder
Gary Snyder (born May 8, 1930) is an American poet, essayist, lecturer, and environmental activist. His early poetry has been associated with the Beat Generation and the San Francisco Renaissance and he has been described as the "poet laureate of Deep Ecology". Snyder is a winner of a Pulitzer Prize for Poetry and the American Book Award. His work, in his various roles, reflects an immersion in both Buddhist spirituality and nature. He has translated literature into English from ancient Chinese and modern Japanese. For many years, Snyder was an academic at the University of California, Davis and for a time served as a member of the California Arts Council. Life and career Early life Gary Sherman Snyder was born in San Francisco, California, to Harold and Lois Hennessy Snyder. Snyder is of German, Scottish, Irish and English ancestry. His family, impoverished by the Great Depression, moved to King County, Washington, when he was two years old. There, they tended dairy-cows, kept laying-hens, had a small orchard, and made cedar-wood shingles. At the age of seven, Snyder was laid up for four months by an accident. "So my folks brought me piles of books from the Seattle Public Library," he recalled in an interview, "and it was then I really learned to read and from that time on was voracious — I figure that accident changed my life. At the end of four months, I had read more than most kids do by the time they're eighteen. And I didn't stop." Also during his ten childhood years in Washington, Snyder became aware of the presence of the Coast Salish people and developed an interest in the Native American peoples in general and their traditional relationship with nature. In 1942, following his parents' divorce, Snyder moved to Portland, Oregon, with his mother and his younger sister, Anthea. Their mother, Lois Snyder Hennessy (born Wilkey), worked during this period as a reporter for The Oregonian. One of his boyhood jobs was as a newspaper copy-boy at the Oregonian. During his teen years, he attended Lincoln High School, worked as a camp counselor, and went mountain-climbing with the Mazamas youth-group. Climbing remained an interest of his, especially during his twenties and thirties. In 1947, he started attending Reed College on a scholarship. Here, he met, and, for a time, roomed with the writer Carl Proujan, and became acquainted with the young poets Philip Whalen and Lew Welch. During his time at Reed, Snyder published his first poems, in a student journal. In 1948 he spent the summer working as a seaman. To get this job, he joined the now-defunct Marine Cooks and Stewards union, and would later work as a seaman in the mid-1950s to gain experience of other cultures in port cities. Snyder married Alison Gass in 1950; they separated after seven months, and divorced in 1952. While attending Reed, Snyder conducted folklore research on the Warm Springs Indian Reservation in central Oregon. He graduated with a dual degree in anthropology and literature in 1951. Snyder's senior thesis, entitled The Dimensions of a Myth, employed perspectives from anthropology, folklore, psychology, and literature to examine a myth of the Pacific Northwest's Haida people. He spent the following few summers working as a timber scaler at Warm Springs, developing relationships with its people that were rooted less in academia. This experience formed the basis for some of his earliest published poems (including "A Berry Feast"), later collected in the book The Back Country. He also encountered the basic ideas of Buddhism and, through its arts, some East Asian traditional attitudes toward nature. He went to Indiana University with a graduate fellowship to study anthropology. (Snyder also began practicing self-taught Zen meditation.) He left after a single semester to return to San Francisco and to 'sink or swim as a poet'. Snyder worked for two summers in the North Cascades in Washington as a fire lookout, on Crater Mountain in 1952 and Sourdough Mountain in 1953 (both locations on the upper Skagit River). His attempts to get another lookout stint in 1954 (at the peak of McCarthyism), however, failed. He found himself barred from working for the government due to his association with the Marine Cooks and Stewards. Instead, he went back to Warm Springs to work in logging as a choker setter. This experience contributed to his Myths and Texts and the essay Ancient Forests of the Far West. The Beats Back in San Francisco, Snyder lived with Whalen, who shared his growing interest in Zen. Snyder's reading of the writings of D. T. Suzuki had in fact been a factor in his decision not to continue as a graduate student in anthropology, and in 1953 he enrolled at the University of California, Berkeley to study Asian culture and languages. He studied ink and wash painting under Chiura Obata and Tang dynasty poetry under Ch'en Shih-hsiang. Snyder continued to spend summers working in the forests, including one summer as a trail-builder in Yosemite. He spent some months in 1955 and 1956 living in a cabin (which he dubbed "Marin-an") outside Mill Valley, California with Jack Kerouac. It was also at this time that Snyder was an occasional student at the American Academy of Asian Studies, where Saburo Hasegawa and Alan Watts, among others, were teaching. Hasegawa introduced Snyder to the treatment of landscape painting as a meditative practice. This inspired Snyder to attempt something equivalent in poetry, and with Hasegawa's encouragement, he began work on Mountains and Rivers Without End, which would be completed and published 40 years later. During these years, Snyder was writing and collecting his own work, as well as embarking on the translation of the "Cold Mountain" poems by the 8th-century Chinese recluse Han Shan; this work appeared in chapbook form in 1959, under the title Riprap & Cold Mountain Poems. Snyder met Allen Ginsberg when the latter sought Snyder out on the recommendation of Kenneth Rexroth. Then, through Ginsberg, Snyder and Kerouac came to know each other. This period provided the materials for Kerouac's novel The Dharma Bums, and Snyder was the inspiration for the novel's main character, Japhy Ryder, in the same way Neal Cassady had inspired Dean Moriarty in On the Road. As the large majority of people in the Beat movement had urban backgrounds, writers like Ginsberg and Kerouac found Snyder, with his backcountry and manual-labor experience and interest in things rural, a refreshing and almost exotic individual. Lawrence Ferlinghetti later referred to Snyder as 'the Thoreau of the Beat Generation'. Snyder read his poem "A Berry Feast" at the poetry reading at the Six Gallery in San Francisco (October 7, 1955) that heard the first reading of Ginsberg's poem "Howl" and marked the emergence into mainstream publicity of the Beats. This also marked Snyder's first involvement with the Beats, although he was not a member of the original New York circle, having entered the scene through his association with Whalen and Welch. As recounted in Kerouac's Dharma Bums, even at age 25 Snyder felt he could have a role in the fateful future meeting of West and East. Snyder's first book, Riprap, which drew on his experiences as a forest lookout and on the trail crew in Yosemite, was published in 1959. Japan and India Independently, some of the Beats, including Whalen, had become interested in Zen, but Snyder was one of the more serious scholars of the subject among them, preparing in every way he could think of for eventual study in Japan. In 1955, the First Zen Institute of America offered him a scholarship for a year of Zen training in Japan, but the State Department refused to issue him a passport, informing him that "it has been alleged you are a Communist." A subsequent District of Columbia Court of Appeals ruling forced a change in policy, and Snyder got his passport. In the end, his expenses were paid by Ruth Fuller Sasaki, for whom he was supposed to work; but initially he served as personal attendant and English tutor to Zen abbot Miura Isshu, at Rinko-in, a temple in Shokoku-ji in Kyoto, where American Buddhist popularizer Dwight Goddard and British author and Japanese culture devotee R. H. Blyth had preceded him. Mornings, after zazen, sutta chanting, and chores for the abbot, he took Japanese classes, bringing his spoken Japanese up to a level sufficient for kōan study. He developed a friendship with Philip Yampolsky an eminent translator and scholar of Zen Buddhism, who took him around Kyoto. In early July 1955, he took refuge and requested to become Miura's disciple, thus formally becoming a Buddhist. In 1958, he returned to California via the Persian Gulf, Turkey, Sri Lanka and various Pacific Islands, voyaging as a crewman in the engine room on the oil tanker Sappa Creek, and took up residence at Marin-an again. He turned one room into a zendo, with about six regular participants. In early June, he met the poet Joanne Kyger. She became his girlfriend, and eventually his wife. In 1959, he shipped for Japan again, where he rented a cottage outside Kyoto. He became the first foreign disciple of Rinzai Rōshi Oda Sesso, the new abbot of Daitoku-ji. He married Kyger on February 28, 1960, immediately after her arrival in Japan, which Fuller Sasaki insisted they do, if they were to live together and be associated with the Nichibei Daiichi Zen Kyokai,. Snyder and Kyger were married from 1960 to 1965. During the period between 1956 and 1969, Snyder went back and forth between California and Japan, studying Zen, working on translations with Fuller Sasaki, and finally living for a while with a group of other people on the small, volcanic island of Suwanosejima. His previous study of written Chinese assisted his immersion in the Zen tradition, which has its roots in Tang Dynasty China, and enabled him to support himself while he was living in Japan. Snyder received the Zen precepts and his dharma name of Chofu ("Listen to the Wind"), and lived occasionally as a de facto monk, but never registered to become a priest, planning eventually to return to the United States to "turn the wheel of the dharma". During this time, he published two collections of his poems from the early to mid 1950s, Myths & Texts (1960), and Six Sections from Mountains and Rivers Without End (1965). This last was the beginning of a project that he was to continue working on until the late 1990s. Much of Snyder's poetry expresses experiences, environments, and insights involved with the work he has done for a living: logger, fire-lookout, steam-freighter crew, translator, carpenter, and itinerant poet, among other things. During his years in Japan, Snyder was also initiated into Shugendo, a highly syncretic ascetic religious cult. In the early 1960s he traveled for six months through India with Kyger, Ginsberg, and Ginsberg's partner, the poet and actor Peter Orlovsky. Snyder and Kyger separated soon after one trip to India, and divorced in 1965. Dharma Bums In the 1950s, Snyder took part in the rise of a strand of Buddhist anarchism emerging from the Beat movement. Snyder was the inspiration for the Japhy Ryder character in Kerouac's novel The Dharma Bums (1958). Snyder had spent considerable time in Japan studying Zen Buddhism, and in 1961 published an essay, "Buddhist Anarchism", where he described the connection he saw between these two traditions, originating in different parts of the world: "The mercy of the West has been social revolution; the mercy of the East has been individual insight into the basic self/void." He advocated "using such means as civil disobedience, outspoken criticism, protest, pacifism, voluntary poverty and even gentle violence" and defended "the right of individuals to smoke ganja, eat peyote, be polygynous, polyandrous or homosexual" which he saw as being banned by "the Judaeo-Capitalist-Christian-Marxist West". Kitkitdizze In 1966, Snyder joined Allen Ginsberg, Richard Baker, future Roshi of the San Francisco Zen Center, and Kriyananda aka Donald J Walters, to buy in the San Juan Ridge area of the Sierra-Nevada foothills, north of Nevada City, Northern California. In 1967 Snyder's book The Back Country appeared, again mainly a collection of poems stretching back over about fifteen years. Snyder devoted a section at the end of the book to his translations of eighteen poems by Kenji Miyazawa. In 1970, Kitkitdizze (as he named his portion of the an Juan Ridge property) would become his home. By that point, Snyder had already spent the summers of 1967 and 1968 with a group of Japanese back-to-the-land drop-outs known as "the Tribe" on Suwanosejima (a small Japanese island in the East China Sea), where they combed the beaches, gathered edible plants, and fished. On the island, on August 6, 1967, he married Masa Uehara, whom he had met in Osaka a year earlier. In 1968, they moved to California with their infant son, Kai (born April 1968). Their second son, Gen, was born a year later. They were shortly able to move onto the San Juan Ridge property, near the South Yuba River, where they and friends built a house that drew on rural-Japanese and Native-American architectural ideas. Later life and writings Regarding Wave appeared in January 1970, a stylistic departure offering poems that were more emotional, metaphoric, and lyrical. From the late 1960s, the content of Snyder's poetry increasingly had to do with family, friends, and community. He continued to publish poetry throughout the 1970s, much of it reflecting his re-immersion in life on the American continent and his involvement in the back-to-the-land movement in the Sierra foothills. His 1974 book Turtle Island, titled after a Native American name for the North American continent, won a Pulitzer Prize. It also influenced numerous West Coast Generation X writers, including Alex Steffen, Bruce Barcott and Mark Morford. His 1983 book Axe Handles, won an American Book Award. Snyder wrote numerous essays setting forth his views on poetry, culture, social experimentation, and the environment. Many of these were collected in Earth House Hold (1969), The Old Ways (1977), The Real Work (1980), The Practice of the Wild (1990), A Place in Space (1995), and The Gary Snyder Reader (1999). In 1979, Snyder published He Who Hunted Birds in His Father's Village: The Dimensions of a Haida Myth, based on his Reed thesis. Snyder's journals from his travel in India in the mid-1960s appeared in 1983 under the title Passage Through India. In these, his wide-ranging interests in cultures, natural history, religions, social critique, contemporary America, and hands-on aspects of rural life, as well as his ideas on literature, were given full-blown articulation. In 1986, Snyder became a professor in the writing program at the University of California, Davis. Snyder is now professor emeritus of English. Snyder was married to Uehara for twenty-two years; the couple divorced in 1989. Snyder married Carole Lynn Koda (October 3, 1947 – June 29, 2006), who would write Homegrown: Thirteen brothers and sisters, a century in America, in 1991, and remained married to her until her death of cancer. She had been born in the third generation of a successful Japanese-American farming family, noted for its excellent rice. She shared Buddhism, extensive travels, and work with Snyder, and performed independent work as a naturalist. As Snyder's involvement in environmental issues and his teaching grew, he seemed to move away from poetry for much of the 1980s and early 1990s. However, in 1996 he published the complete Mountains and Rivers Without End, a mixture of the lyrical and epic modes celebrating the act of inhabitation on a specific place on the planet. This work was written over a 40-year period. It has been translated into Japanese, French and Russian. In 2004 Snyder published Danger on Peaks, his first collection of new poems in twenty years. Snyder was awarded the Levinson Prize from the journal Poetry, the American Poetry Society Shelley Memorial Award (1986), was inducted into the American Academy of Arts and Letters (1987), and won the 1997 Bollingen Prize for Poetry and, that same year, the John Hay Award for Nature Writing. Snyder also has the distinction of being the first American to receive the Buddhism Transmission Award (for 1998) from the Japan-based Bukkyo Dendo Kyokai Foundation. For his ecological and social activism, Snyder was named as one of the 100 visionaries selected in 1995 by Utne Reader. Snyder's life and work was celebrated in John J. Healy's 2010 documentary The Practice of the Wild. The film, which debuted at the 53rd San Francisco International Film Festival, features wide-ranging, running conversations between Snyder and poet, writer and longtime colleague Jim Harrison, filmed mostly on the Hearst Ranch in San Simeon, California. The film also shows archival photographs and film of Snyder's life. Work Poetics Gary Snyder uses mainly common speech-patterns as the basis for his lines, though his style has been noted for its "flexibility" and the variety of different forms his poems have taken. He typically uses neither conventional meters nor intentional rhyme. "Love and respect for the primitive tribe, honour accorded the Earth, the escape from city and industry into both the past and the possible, contemplation, the communal", such, according to Glyn Maxwell, is the awareness and commitment behind the specific poems. The author and editor Stewart Brand once wrote: "Gary Snyder's poetry addresses the life-planet identification with unusual simplicity of style and complexity of effect." According to Jody Norton, this simplicity and complexity derives from Snyder's use of natural imagery (geographical formations, flora, and fauna) in his poems. Such imagery can be both sensual at a personal level yet universal and generic in nature. In the 1968 poem "Beneath My Hand and Eye the Distant Hills, Your Body," the author compares the intimate experience of a lover's caress with the mountains, hills, cinder cones, and craters of the Uintah Mountains. Readers become explorers on both a very private level as well as a very public and grand level. A simplistic touch becoming a very complex interaction occurring at multiple levels. This is the effect Snyder intended. In an interview with Faas, he states, "There is a direction which is very beautiful, and that's the direction of the organism being less and less locked into itself, less and less locked into its own body structure and its relatively inadequate sense organs, towards a state where the organism can actually go out from itself and share itself with others." Snyder has always maintained that his personal sensibility arose from his interest in Native Americans and their involvement with nature and knowledge of it; indeed, their ways seemed to resonate with his own. And he has sought something akin to this through Buddhist practices, Yamabushi initiation, and other experiences and involvements. However, since his youth he has been quite literate, and he has written about his appreciation of writers of similar sensibilities, like D. H. Lawrence, William Butler Yeats, and some of the great ancient Chinese poets. William Carlos Williams was another influence, especially on Snyder's earliest published work. Starting in high school, Snyder read and loved the work of Robinson Jeffers, his predecessor in poetry of the landscape of the American West; but, whereas Jeffers valued nature over humankind, Snyder saw humankind as part of nature. Snyder commented in interviews, "I have some concerns that I'm continually investigating that tie together biology, mysticism, prehistory, general systems theory". Snyder argues that poets, and humans in general, need to adjust to very long timescales, especially when judging the consequences of their actions. His poetry examines the gap between nature and culture so as to point to ways in which the two can be more closely integrated. In 2004, receiving the Masaoka Shiki International Haiku Awards Grand Prize, Snyder highlighted traditional ballads and folk songs, Native American songs and poems, William Blake, Walt Whitman, Jeffers, Ezra Pound, Noh drama, Zen aphorisms, Federico García Lorca, and Robert Duncan as significant influences on his poetry, but added, "the influence from haiku and from the Chinese is, I think, the deepest." Romanticism Snyder is among those writers who have sought to dis-entrench conventional thinking about primitive peoples that has viewed them as simple-minded, ignorantly superstitious, brutish, and prone to violent emotionalism. In the 1960s Snyder developed a "neo-tribalist" view akin to the "post-modernist" theory of French Sociologist Michel Maffesoli. The "re-tribalization" of the modern, mass-society world envisioned by Marshall McLuhan, with all of the ominous, dystopian possibilities that McLuhan warned of, subsequently accepted by many modern intellectuals, is not the future that Snyder expects or works toward. Snyder's is a positive interpretation of the tribe and of the possible future. Todd Ensign describes Snyder's interpretation as blending ancient tribal beliefs and traditions, philosophy, physicality, and nature with politics to create his own form of Postmodern environmentalism. Snyder rejects the perspective which portrays nature and humanity in direct opposition to one another. Instead, he chooses to write from multiple viewpoints. He purposely sets out to bring about change on the emotional, physical, and political levels by emphasizing the ecological problems faced by today's society. Beat Gary Snyder is widely regarded as a member of the Beat Generation circle of writers: he was one of the poets that read at the famous Six Gallery event, and was written about in one of Kerouac's most popular novels, The Dharma Bums. Some critics argue that Snyder's connection with the Beats is exaggerated and that he might better be regarded as a part of the San Francisco Renaissance, which developed independently. Snyder himself has some reservations about the label "Beat", but does not appear to have any strong objection to being included in the group. He often talks about the Beats in the first person plural, referring to the group as "we" and "us". A quotation from a 1974 interview at the University of North Dakota Writers Conference (published in The Beat Vision): I never did know exactly what was meant by the term 'The Beats', but let's say that the original meeting, association, comradeship of Allen [Ginsberg], myself, Michael [McClure], Lawrence [Ferlinghetti], Philip Whalen, who's not here, Lew Welch, who's dead, Gregory [Corso], for me, to a somewhat lesser extent (I never knew Gregory as well as the others) did embody a criticism and a vision which we shared in various ways, and then went our own ways for many years. Where we began to come really close together again, in the late '60s, and gradually working toward this point, it seems to me, was when Allen began to take a deep interest in Oriental thought and then in Buddhism which added another dimension to our levels of agreement; and later through Allen's influence, Lawrence began to draw toward that; and from another angle, Michael and I after the lapse of some years of contact, found our heads very much in the same place, and it's very curious and interesting now; and Lawrence went off in a very political direction for a while, which none of us had any objection with, except that wasn't my main focus. It's very interesting that we find ourselves so much on the same ground again, after having explored divergent paths; and find ourselves united on this position of powerful environmental concern, critique of the future of the individual state, and an essentially shared poetics, and only half-stated but in the background very powerfully there, a basic agreement on some Buddhist type psychological views of human nature and human possibilities. Snyder has also commented "The term Beat is better used for a smaller group of writers ... the immediate group around Allen Ginsberg and Jack Kerouac, plus Gregory Corso and a few others. Many of us ... belong together in the category of the San Francisco Renaissance. ... Still, beat can also be defined as a particular state of mind ... and I was in that mind for a while". Bibliography Riprap and Cold Mountain Poems (1959) Myths & Texts (1960) Six Sections from Mountains and Rivers Without End (1965) The Back Country (Fulcrum, 1967) Regarding Wave (1969) Earth House Hold (1969) Smokey the Bear Sutra (1969) Turtle Island (1974) The Old Ways (1977) He Who Hunted Birds in His Father's Village: The Dimensions of a Haida Myth (1979) The Real Work: Interviews & Talks 1964-1979 (1980) Axe Handles (1983) Passage Through India (1983) Left Out in the Rain (1988) The Practice of the Wild (1990) No Nature: New and Selected Poems (1992) A Place in Space (1995) narrator of the audio book version of Kazuaki Tanahashi's Moon in a Dewdrop from Dōgen's Shōbōgenzō Mountains and Rivers Without End (1996) The Geography Of Home (Poetry book)(1999) The Gary Snyder Reader: Prose, Poetry, and Translations (1999) The High Sierra of California, with Tom Killion (2002) Look Out: a Selection of Writings (November 2002) Danger on Peaks (2005) Back on the Fire: Essays (2007) The Selected Letters of Allen Ginsberg and Gary Snyder, 1956-1991 (2009) Tamalpais Walking, with Tom Killion (2009) The Etiquette of Freedom, with Jim Harrison (2010): film by Will Hearst with book edited by Paul Ebenkamp Nobody Home: Writing, Buddhism, and Living in Places, with Julia Martin, Trinity University Press (2014) This Present Moment (April 2015) Distant Neighbors: The Selected Letters of Wendell Berry and Gary Snyder (May 2015) The Great Clod: Notes and Memories on Nature and History in East Asia (March 2016) Dooby Lane: Also Known as Guru Road, A Testament Inscribed in Stone Tablets by DeWayne Williams, with Peter Goin (October 2016) Collected Poems (The Library of America, 2022) ISBN 9781598537215 Citations General sources Charters, Ann (ed.). The Portable Beat Reader. Penguin Books. New York. 1992. (hc); (pbk) Hunt, Anthony. "Genesis, Structure, and Meaning in Gary Snyder's Mountains and Rivers Without End" Univ. of Nevada Press. 2004. Knight, Arthur Winfield. Ed. The Beat Vision (1987) Paragon House. ; (pbk) Kyger, Joanne. Strange Big Moon: The Japan and India Journals: 1960–1964 (2000) North Atlantic Books. Smith, Eric Todd. Reading Gary Snyder's Mountains and Rivers Without End (1999) Boise State University. Snyder, Gary. The Politics of Ethnopoetics (1975) Snyder essay A Place in Space Snyder, Gary. 1980. The Real Work: Interviews & Talks 1964–1979. New Directions, New York. (hbk); (pbk) Stirling, Isabel. Zen Pioneer: The Life & Works of Ruth Fuller Sasaki (2006) Shoemaker & Hoard. Suiter, John. Poets on the Peaks (2002) Counterpoint. ; (pbk) Western Literature Association. Updating the Literary West (1997) Texas Christian University Press. Further reading Sherlock, John. (2010). Gary Snyder: a bibliography of works by and about Gary Snyder . UC Davis Library. External links Profile at Poetry Foundation Profile at Poets.org Snyder talk "Mountains and Rivers without End" at the Smithsonian Museums of Asian art (Audio 1 hr) at 12 July 2008. Talk programme "The Wild Mind Of Gary Snyder" by Trevor Carolan and "Writers and the War Against Nature" by Gary Snyder in Shambhala Sun magazine 2007 Public Access TV interview (Nevada County TeleVision), 61 minutes "Gary Snyder" by Bert Almon from the Western Writers Series Digital Editions at Boise State University New York Times profile "A Poem, 40 Years Long" 6 October 1996 Gary Snyder on Art, Anarchy and the Environment (2010 San Francisco Film Society interview) Gary Snyder Papers at Special Collections Dept., University Library, University of California, Davis Gary Snyder. Letters to Shandel Parks MSS 719. Special Collections & Archives, UC San Diego Library. Records of Gary Snyder are held by Simon Fraser University's Special Collections and Rare Books Western American Literature Journal: Gary Snyder 1930 births Activists from California American Book Award winners American environmentalists American male non-fiction writers American male poets American non-fiction environmental writers American people of English descent American people of German descent American people of Irish descent American people of Scottish descent American spiritual writers American Buddhists American Zen Buddhists Anti-consumerists Beat Generation writers Bollingen Prize recipients Deep ecologists Engaged Buddhists English-language haiku poets Industrial Workers of the World members Lincoln High School (Portland, Oregon) alumni Living people Members of the American Academy of Arts and Letters The Oregonian people PEN Oakland/Josephine Miles Literary Award winners Pulitzer Prize for Poetry winners Reed College alumni Rinzai Buddhists University of California, Davis faculty Writers from Portland, Oregon
12888
https://en.wikipedia.org/wiki/Francis%20Gary%20Powers
Francis Gary Powers
Francis Gary Powers (August 17, 1929 – August 1, 1977) was an American pilot whose Central Intelligence Agency (CIA) Lockheed U-2 spy plane was shot down while flying a reconnaissance mission in Soviet Union airspace, causing the 1960 U-2 incident. He later worked as a helicopter pilot for KNBC in Los Angeles and died in a 1977 helicopter crash. Early life and education Powers was born August 17, 1929, in Jenkins, Kentucky, the son of Oliver Winfield Powers (1904–1970), a coal miner, and his wife Ida Melinda Powers (; 1905–1991). His family eventually moved to Pound, Virginia, just across the state border. He was the second-born and only male of six children. His family lived in a mining town, and because of the hardships associated with living in such a town, his father wanted Powers to become a physician. He hoped his son would achieve the higher earnings of such a profession and felt that this would involve less hardship than any job in his hometown. Education and service After graduating with a bachelor's degree from Milligan College in Tennessee in June 1950, Powers enlisted in the United States Air Force in October. He was commissioned as a second lieutenant in December 1952 after completing his advanced training with USAF Pilot Training Class 52-H at Williams Air Force Base, Arizona. Powers was then assigned to the 468th Strategic Fighter Squadron at Turner Air Force Base, Georgia, as a Republic F-84 Thunderjet pilot. Powers married Barbara Gay Moore in Newnan, Georgia, on April 2, 1955. In January 1956 he was recruited by the CIA. In May 1956 he began U-2 training at Watertown Strip, Nevada. His training was complete by August 1956 and his unit, the Second Weather Observational Squadron (Provisional) or Detachment 10-10, was deployed to Incirlik Air Base, Turkey. By 1960, Powers was already a veteran of many covert aerial reconnaissance missions. Family members believed that he was a NASA weather reconnaissance pilot. The U-2 incident Powers was discharged from the Air Force in 1956 with the rank of captain. He then joined the CIA's U-2 program at the civilian grade of GS-12. U-2 pilots flew espionage missions at altitudes of , supposedly above the reach of Soviet air defenses. The U-2 was equipped with a state-of-the-art camera designed to take high-resolution photos from the stratosphere over hostile countries, including the Soviet Union. U-2 missions systematically photographed military installations and other important sites. Reconnaissance mission The primary mission of the U-2s was to overfly the Soviet Union. Soviet intelligence had been aware of encroaching U-2 flights at least since 1958 if not earlier but lacked effective countermeasures until 1960. On May 1, 1960, Powers' U-2A, 56-6693, departed from a military airbase in Peshawar, Pakistan, with support from the U.S. Air Station at Badaber (Peshawar Air Station). This was to be the first attempt "to fly all the way across the Soviet Union ... but it was considered worth the gamble. The planned route would take us deeper into Russia than we had ever gone, while traversing important targets never before photographed." Shot down Powers was shot down by an S-75 Dvina (SA-2 "Guideline") surface-to-air missile over Sverdlovsk. A total of 14 Dvinas were launched, one of which hit a MiG-19 jet fighter which was sent to intercept the U-2 but could not reach a high enough altitude. Its pilot, Sergei Safronov, ejected but died of his injuries. Another Soviet aircraft, a newly manufactured Su-9 on a transit flight, also attempted to intercept Powers' U-2. The unarmed Su-9 was directed to ram the U-2 but missed because of the large differences in speed. As Powers flew near Kosulino in the Ural Region, three S-75 Dvinas were launched at his U-2, with the first one hitting the aircraft. "What was left of the plane began spinning, only upside down, the nose pointing upward toward the sky, the tail down toward the ground." Powers was unable to activate the camera's self-destruct mechanism before he was thrown out of the plane after releasing the canopy and his seat belt. According to his own book Operation Overflight, it was a misunderstanding that the whole plane could be blown up. While descending under his parachute, Powers had time to scatter his escape map, and rid himself of part of his suicide device, a silver dollar coin suspended around his neck containing a poison-laced injection pin, though he kept the poison pin. "Yet I was still hopeful of escape." He hit the ground hard, was immediately captured, and taken to Lubyanka Prison in Moscow. Powers did note a second chute after landing on the ground, "some distance away and very high, a lone red and white parachute". Attempted deception by the U.S. government When the U.S. government learned of Powers' disappearance over the Soviet Union, they lied that a "weather plane" had strayed off course after its pilot had "difficulties with his oxygen equipment". What CIA officials did not realize was that the plane crashed almost fully intact and that the Soviets had recovered its pilot and the plane's equipment, including its top-secret high-altitude camera. Powers was interrogated extensively by the KGB for months before he made a confession and a public apology for his part in espionage. Portrayal in U.S. media Following admission by the White House that Powers had been captured alive, American media depicted Powers as an all-American pilot hero, who never smoked or touched alcohol. In fact, Powers smoked and drank socially. The CIA urged that his wife Barbara be given sedatives before speaking to the press and gave her talking points that she repeated to the press to portray her as a devoted wife. Her broken leg, according to the CIA disinformation, was the result of a water-skiing accident, when in fact it happened after she had had too much to drink and was dancing with another man. In the course of his trial for espionage in the Soviet Union, Powers confessed to the charges against him and apologized for violating Soviet airspace to spy on the Soviets. In the wake of his apology, American media often depicted Powers as a coward and even as a symptom of the decay of America's "moral character." Pilot testimony compromised by newspaper reports Powers tried to limit the information he shared with the KGB to that which could be determined from the remains of his plane's wreckage. He was hampered by information appearing in the western press. A KGB major stated "there's no reason for you to withhold information. We'll find it out anyway. Your Press will give it to us." However, he limited his divulging of CIA contacts to one individual, with a pseudonym of "Collins". At the same time, he repeatedly stated the maximum altitude for the U-2 was , significantly lower than its actual flight ceiling. Political consequence The incident set back talks between Khrushchev and Eisenhower. Powers' interrogations ended on June 30, and his solitary confinement ended on July 9. On August 17, 1960, his trial began for espionage before the military division of the Supreme Court of the USSR. Lieutenant General Borisoglebsky, Major General Vorobyev, and Major General Zakharov presided. Roman Rudenko acted as prosecutor in his capacity of Procurator General of the Soviet Union. Mikhail I. Grinev served as Powers' defense counsel. In attendance were his parents and sister, and his wife Barbara and her mother. His father brought along his attorney Carl McAfee, while the CIA provided two additional attorneys. Conviction On August 19, 1960, Powers was convicted of espionage, "a grave crime covered by Article 2 of the Soviet Union's law 'On Criminality Responsibility for State Crimes'". His sentence consisted of 10 years' confinement, three of which were to be in a prison, with the remainder in a labor camp. The US Embassy "News Bulletin" stated, according to Powers, "as far as the government was concerned, I had acted in accordance with the instructions given me and would receive my full salary while imprisoned". He was held in Vladimir Central Prison, about east of Moscow, in building number 2 from September 9, 1960, until February 8, 1962. His cellmate was Zigurds Krūmiņš, a Latvian political prisoner. Powers kept a diary and a journal while confined. Additionally, he learned carpet weaving from his cellmate to pass the time. He could send and receive a limited number of letters to and from his family. The prison now contains a small museum with an exhibit on Powers, who allegedly developed a good rapport with Soviet prisoners there. Some pieces of the plane and Powers' uniform are on display at the Monino Airbase museum near Moscow. Prisoner exchange CIA opposition to exchange The CIA, in particular, chief of CIA Counterintelligence James Jesus Angleton, opposed exchanging Powers for Soviet KGB Colonel William Fisher, known as "Rudolf Abel", who had been caught by the FBI and tried and jailed for espionage. First, Angleton believed that Powers might have deliberately defected to the Soviet side. CIA documents released in 2010 indicate that U.S. officials did not believe Powers' account of the incident at the time, because it was contradicted by a classified National Security Agency (NSA) report which alleged that the U-2 had descended from before changing course and disappearing from radar. The NSA report remains classified as of 2022. In any event, Angleton suspected that Powers had already revealed all he knew to the Soviets and therefore reasoned that Powers was worthless to the U.S. On the other hand, according to Angleton, William Fisher had revealed little to the CIA, refusing to disclose even his real name, and for this reason, William Fisher was still of potential value. However, Barbara Powers, Gary Powers' wife, was allegedly often drinking and having affairs. On June 22, 1961, she was pulled over by the police after driving erratically and was caught driving under the influence. To avoid bad publicity for the wife of the well-known CIA operative, doctors tasked by the CIA to keep Barbara out of the limelight arranged to have her committed to a psychiatric ward in Augusta, Georgia, under strict supervision. She was eventually released to the care of her mother. However, the CIA feared that Gary Powers languishing in Soviet prison might learn of Barbara's plight and as a result reach a state of desperation causing him to reveal to the Soviets whatever secrets he had not already revealed. Thus, Barbara unwittingly may have aided the cause of the approval of the prisoner exchange involving her husband and William Fisher. Angleton and others at the CIA still opposed the exchange but President John F. Kennedy approved it. The exchange On February 10, 1962, Powers was exchanged, along with U.S. student Frederic Pryor, for Soviet KGB Colonel Rudolf Abel. Due to political differences between the Soviet Union and the German Democratic Republic at the time, Pryor was turned over to American authorities at Checkpoint Charlie, before the exchange of Powers for Fisher was allowed to proceed on the Glienicke Bridge. Powers credited his father with the swap idea. When released, Powers' total time in captivity was 1 year, 9 months, and 10 days. Aftermath Powers initially received a cold reception on his return home. He was criticized for not activating his aircraft's self-destruct charge to destroy the camera, photographic film, and related classified parts. He was also criticized for not using a CIA-issued "suicide pill" to kill himself (a coin with shellfish toxin embedded in its grooves, revealed during CIA testimony to the Church Committee in 1975). He was debriefed extensively by the CIA, Lockheed Corporation, and the Air Force, after which a statement was issued by CIA director John McCone that "Mr. Powers lived up to the terms of his employment and instructions in connection with his mission and in his obligations as an American." On March 6, 1962, he appeared before a Senate Armed Services Select Committee hearing chaired by Senator Richard Russell Jr. which included Senators Prescott Bush, Leverett Saltonstall, Robert Byrd, Margaret Chase Smith, John Stennis, Strom Thurmond, and Barry Goldwater. During the hearing, Senator Saltonstall stated, "I commend you as a courageous, fine young American citizen who lived up to your instructions and who did the best you could under very difficult circumstances." Senator Bush declared, "I am satisfied he has conducted himself in exemplary fashion and in accordance with the highest traditions of service to one's country, and I congratulate him upon his conduct in captivity." Senator Goldwater sent him a handwritten note: "You did a good job for your country." Divorce and remarriage Powers sued his wife for divorce on August 14, 1962, claiming she cursed and abused him for no reason. Powers stated that the reasons for the divorce included her infidelity and alcoholism, adding that she constantly threw tantrums and overdosed on pills shortly after his return. He started a relationship with Claudia Edwards "Sue" Downey, whom he had met while working briefly at CIA Headquarters. Downey had a child, Dee, from her previous marriage. They were married on October 26, 1963. Their son Francis Gary Powers Jr. was born on June 5, 1965. The marriage proved to be a very happy one, and Sue worked hard to preserve her husband's legacy after his death. Praise During a speech in March 1964, former CIA Director Allen Dulles said of Powers, "He performed his duty in a very dangerous mission and he performed it well, and I think I know more about that than some of his detractors and critics know, and I am glad to say that to him tonight." Later career Powers worked for Lockheed as a test pilot from 1962 to 1970, though the CIA paid his salary. In 1970, he wrote the book Operation Overflight with co-author Curt Gentry. Lockheed fired him, because "the book's publication had ruffled some feathers at Langley." Powers then became a traffic reporting airplane pilot for Los Angeles radio station KGIL. After that he became a helicopter news reporter for KNBC television. Death Powers was piloting a helicopter for Los Angeles TV station KNBC Channel 4 over the San Fernando Valley on August 1, 1977, when the aircraft crashed, killing him and his cameraman George Spears. They had been recording video following brush fires in Santa Barbara County in the KNBC helicopter and were heading back when the crash occurred. His Bell 206 JetRanger helicopter ran out of fuel and crashed at the Sepulveda Dam recreational area in Encino, California, several miles short of its intended landing site at Burbank Airport. The National Transportation Safety Board report attributed the probable cause of the crash to pilot error. According to Powers' son, an aviation mechanic had repaired a faulty fuel gauge without informing Powers, who subsequently misread it. At the last moment, it is surmised that he noticed children playing in the area and directed the helicopter elsewhere to avoid landing on them. He might have landed safely if not for the last-second deviation, which compromised his autorotative descent. Powers was survived by his wife, children (Claudia Dee and Francis Gary Powers Jr.), and five sisters. He is buried in Arlington National Cemetery as an Air Force veteran. Honors Powers received the CIA's Intelligence Star in 1972 after his return from the Soviet Union. Powers was originally scheduled to receive it in 1963 along with other pilots involved in the CIA's U-2 program, but the award was postponed for political reasons. In 1970, Powers published his first—and only—book review, on a work about aerial reconnaissance, Unarmed and Unafraid by Glenn Infield, in the monthly magazine Business & Commercial Aviation. "The subject has great interest to me," he said, in submitting his review. In 1998, newly declassified information revealed that Powers' mission had been a joint USAF/CIA operation. In 2000, on the 40th anniversary of the U-2 Incident, his family was presented with his posthumously awarded Prisoner of War Medal, Distinguished Flying Cross, and National Defense Service Medal. In addition, CIA Director George Tenet authorized Powers to posthumously receive the CIA's coveted Director's Medal for extreme fidelity and extraordinary courage in the line of duty. On June 15, 2012, Powers was posthumously awarded the Silver Star medal for "demonstrating 'exceptional loyalty' while enduring harsh interrogation in the Lubyanka Prison in Moscow for almost two years." Air Force Chief of Staff General Norton Schwartz presented the decoration to Powers' grandchildren, Trey Powers, 9, and Lindsey Berry, 29, in a Pentagon ceremony. Legacy Powers' son, Francis Gary Powers Jr., founded the Cold War Museum in 1996. Affiliated with the Smithsonian Institution, it was essentially a traveling exhibit until it found a permanent home in 2011 on a former Army communications base outside Washington, D.C. In popular culture In the 1976 telemovie Francis Gary Powers: The True Story of the U-2 Spy Incident, Powers was played by Lee Majors. In 1999, the History Channel aired Mystery of the U2, hosted by Arthur Kent as part of their History Undercover series. The program was produced by Indigo Films. In the 2015 movie Bridge of Spies, dramatizing the negotiations to repatriate Powers, he is portrayed by Austin Stowell, with Tom Hanks starring as negotiator James Donovan. In April 2018, The Aviationist featured an article about the song "Powers Down", a tribute to Powers. References Citations Bibliography Powers, Francis Gary with Gentry, Curt. Operation Overflight. Hodder & Stoughton Ltd, 1971 (hard cover) . Notes Further reading External links Family webpage on Gary Powers Info on The Cold War Museum CIA FOIA documents on Gary Powers FBI Records: The Vault – Francis Gary Powers Documents and Photographs regarding the U-2 Spy Plane Incident of 1960, Dwight D. Eisenhower Presidential Library Check-Six.com The Francis Gary Powers Helo Crash Transcripts of the Soviet court trial 1962 Russia frees US spy plane pilot View images of the Francis Gary Powers U-2 Pilot Cinderella stamps on the artist's webSite. IMDB page for the 1976 TV movie. ANC Explorer 1929 births 1977 deaths American people convicted of spying for the United States American people imprisoned abroad American people imprisoned in the Soviet Union American prisoners of war American spies against the Soviet Union Aviators killed in aviation accidents or incidents in the United States Burials at Arlington National Cemetery Films about shot-down aviators Inmates of Vladimir Central Prison Lockheed people Milligan University alumni Military personnel from Kentucky Military personnel from Virginia People from Jenkins, Kentucky People from Pound, Virginia People of the Central Intelligence Agency Recipients of the Distinguished Flying Cross (United States) Recipients of the Intelligence Star Recipients of the Silver Star Reconnaissance pilots Shot-down aviators United States Air Force officers Victims of aviation accidents or incidents in 1977 Victims of helicopter accidents or incidents in the United States
13037
https://en.wikipedia.org/wiki/Guantanamo%20Bay%20Naval%20Base
Guantanamo Bay Naval Base
Guantanamo Bay Naval Base (), officially known as Naval Station Guantanamo Bay or NSGB, (also called GTMO, pronounced Gitmo as jargon by members of the U.S. military) is a United States military base located on of land and water on the shore of Guantánamo Bay at the southeastern end of Cuba. It has been permanently leased to the United States since 1903 as a coaling station and naval base, making it the oldest overseas U.S. naval base in the world. The lease was $2,000 in gold per year until 1934, when the payment was set to match the value in gold in dollars; in 1974, the yearly lease was set to $4,085. Since taking power in 1959, the Cuban communist government has consistently protested against the U.S. presence on Cuban soil, arguing that the base "was imposed on Cuba by force" and is "illegal under international law." Since 2002, the naval base has contained a military prison, for alleged unlawful combatants captured in Afghanistan, Iraq, and other places during the War on Terror. Cases of alleged torture of prisoners by the U.S. military, and their denial of protection under the Geneva Conventions, have been criticized. The 1903 lease has no fixed expiration date, and as such it can only be ended if the US Navy decided to abandon the area or both countries agreed mutually to end the lease. Geography The Naval Base occupies three main geographical sections: Leeward Point, Windward Point, and Guantánamo Bay. Guantánamo Bay physically divides the Naval Station into sections. The bay extends past the boundaries of the base into Cuba, where the bay is then referred to as Bahía de Guantánamo. Guantánamo Bay contains several cays, which are identified as Hospital Cay, Medico Cay, North Toro Cay, and South Toro Cay. Leeward Point of the Naval Station is the site of the active airfield. Major geographical features on Leeward Point include Mohomilla Bay and the Guantánamo River. Three beaches exist on the Leeward side. Two are available for use by base residents, while the third, Hicacal Beach, is closed. Windward Point contains most of the activities at the Naval Station. There are nine beaches available to base personnel. The highest point on the base is John Paul Jones Hill (named after the Revolutionary War hero of the same name) at a total of . The geography of Windward Point is such that there are many coves and peninsulas along the bay shoreline providing ideal areas for mooring ships. According to Stephen Benz, Stephen Crane described the area as "craggy" and "cut with ravines." He also described the coastline as covered in ridges and "chalky cliffs". History Spanish era The area surrounding Guantanamo Bay was originally inhabited by the Taíno people. On 30 April 1494, Christopher Columbus, on his second voyage, arrived and spent the night. The place where Columbus landed is now known as Fisherman's Point. Columbus declared the bay Puerto Grande. The bay and surrounding areas briefly came under British control during the War of Jenkins' Ear. Prior to the outbreak of the conflict, the bay was referred to as Walthenham Harbor by British cartographers. The British expeditionary force renamed the bay Cumberland Bay. They eventually retreated from the area after an attempt to march to Santiago de Cuba was repulsed by Spanish troops. Spanish–American War During the Spanish–American War, the U.S. fleet attacking Santiago secured Guantánamo's harbor for protection during the hurricane season of 1898. The Marines landed at Guantanamo Bay with naval support, and American and Cuban forces routed the defending Spanish troops. There is a monument on McCalla Hill to one Navy officer and five Marines who died in battle at Guantanamo Bay. The war ended with the Treaty of Paris of 1898, in which Spain formally relinquished control of Cuba. Although the war was over, the United States maintained a strong military presence on the island. In 1901 the United States government passed the Platt Amendment as part of an Army Appropriations Bill. Section VII of this amendment read After initial resistance by the Cuban Constitutional Convention, the Platt Amendment was incorporated into the Constitution of the Republic of Cuba in 1901. The Constitution took effect in 1902, and land for a naval base at Guantanamo Bay was granted to the United States the following year. Permanent lease The 1903 lease agreement, which has no fixed expiration date, was executed in two parts. The first, signed in February, consisted of the following provisions: Agreement – The United States of America and the Republic of Cuba, desiring to maintain the Independence of Cuba, will enter into a lease for lands necessary for US Naval Stations. Article 1 – Describes the boundaries of the areas being leased, Guantanamo Bay and Bahia Honda. Article 2 – The U.S. may occupy, use, and modify the properties to fit the needs of a coaling and naval station, only. Vessels in the Cuban trade shall have free passage. Article 3 – Cuba retains ultimate sovereignty, but during the occupation, the U.S. exercises sole jurisdiction over the areas described in Article 1. Under conditions to be agreed on, the U.S. has the right to acquire, by purchase or eminent domain, any land included therein. The second part, signed five months later in July 1903, consisted of the following provisions: Article 1 – Payment is $2000 gold coin, annually. All private lands within the boundaries shall be acquired by Cuba. The U.S. will advance rental payments to Cuba to facilitate those purchases. Article 2 – The U.S. shall pay for a survey of the sites and mark the boundaries with fences. Article 3 – There will be no commercial or other enterprise within the leased areas. Article 4 – Mutual extradition Article 5 – Not ports of entry. Article 6 – Ships shall be subject to Cuban port police. The U.S. will not obstruct entry or departure into the bay. Article 7 – This proposal is open for seven months. SIGNED Theodore Roosevelt and Jose M Garcia Montes. In 1934, the United States unilaterally changed the payment from gold coin to U.S. dollars per the Gold Reserve Act. The lease amount was set at US$3,386.25, based on the price of gold at the time. In 1973, the U.S. adjusted the lease amount to $3,676.50, and in 1974 to $4,085, based on further increases to the price of gold in USD. Payments have been sent annually, but only one lease payment has been accepted since the Cuban Revolution and Fidel Castro claimed that this check was deposited due to confusion in 1959. The Cuban government has not deposited any other lease check since that time. Before and during World War II The Naval Base became a prime area for winter training for the Atlantic fleet. Due to the large increase in population during the training months, the Naval Base quickly established facilities to support everyday functions. Guantanamo Bay fleet range system began its construction as early as 1905 under Captain Rogers. Four primary ranges were built during its construction: Range Alpha, Range Beta, Range Charlie, and Range Delta. Due to the size of the base, sections were shut down in order for range operations to proceed. Ironically, the concrete range system's size and cost led to its downfall. Strong evidence suggests the fleet ranges were used throughout World War I and the beginning of World War II. During World War II, the base was set up to use a nondescript number for postal operations. The base used the Fleet Post Office, Atlantic, in New York City, with the address: 115 FPO NY. The base was also an important intermediate distribution point for merchant shipping convoys from New York City and Key West, Florida, to the Panama Canal and the islands of Puerto Rico, Jamaica, and Trinidad and Tobago. Cold War and beyond From 1953 to 1959, thousands of Cubans commuted daily from outside the base to jobs within it. In mid-1958, vehicular traffic was stopped; workers were required to walk through the base's several gates. Public Works Center buses were pressed into service almost overnight to carry the tides of workers to and from the gate. The "Cactus Curtain" is a term describing the line separating the naval base from Cuban-controlled territory, an allusion to Europe's Iron Curtain, the Bamboo Curtain in East Asia, and the similar Ice Curtain in the Bering Strait. After the Cuban Revolution, some Cubans sought refuge on the Guantanamo Bay Naval Base. In late 1961, Cuban troops planted an long, wide barrier of Opuntia cactus along the northeastern section of the fence surrounding the base in order to prevent checkpoint evasion when moving between the base and Cuba proper. The curtain forms part of a "no man's land" that encircles the base. This area is complete with perimeter patrols, outposts featuring sandbags, and watchtowers, and has been complemented with barbed wire fences, minefields, and cacti. Apart from the cacti, both U.S. and Cuban troops erected, maintained, and otherwise manned these defenses, primarily to prevent checkpoint evasion and possible invasion from the other side. U.S. and Cuban troops placed some 55,000 anti-personnel and anti-tank land mines across the "no man's land" around the perimeter of the naval base, creating the second-largest minefield in the world, and the largest in the Western Hemisphere. Initially, the mines were laid down by US troops, who also laid down signs stating that the landmines were "precautions" and should not be viewed as "aggressive". In response, Cuban troops also laid down their own mines, with both sides completing their minefields in 1961. Between 1961 and 1965, landmine explosions resulted in the deaths of at least 10 people, including as the result of engineering accidents and late-night partygoers. On 16 May 1996, U.S. President Bill Clinton ordered the demining of the American field. They have since been replaced with motion and sound sensors to detect intruders on the base. The Cuban government has not removed its corresponding minefield outside the perimeter. During the Cuban Missile Crisis in 1962, the families of military personnel were evacuated from the base. Notified of the evacuation on 22 October, evacuees were told to pack one suitcase per family member, to bring evacuation and immunization cards, to tie pets in the yard, to leave the keys to the house on the dining table, and to wait in front of the house for buses. Dependents travelled to the airfield for flights to the United States, or to ports for passage aboard evacuation ships. After the crisis was resolved, family members were allowed to return to the base in December 1962. From 1939, the base's water was supplied by pipelines that drew water from the Yateras River about northeast of the base. The U.S. government paid a fee for this; in 1964, it was about $14,000 a month for about per day. In 1964, the Cuban government stopped the flow. The base had about of water in storage, and strict water conservation was put into effect immediately. The U.S. first imported water from Jamaica by barge, then relocated a desalination plant from San Diego (Point Loma). When the Cuban government accused the United States of "stealing water," base commander John D. Bulkeley ordered that the pipelines be cut and a section removed. A length of the diameter pipe and a length of the diameter pipe were lifted from the ground and the openings sealed. During the 1960s and 1970s, the base had problems with alcohol and racial tension. Harassment and strip searches also became a regular occurrence for Cuban workers on the base. Several old guns from the USS Monongahela (1862) have been salvaged and placed around the base. The old warship served as a storeship at Guantanamo until destroyed by fire on 17 March 1908. A gun was salvaged from her wreck and put on display at the Naval Station. Since the gun was deformed by the heat from the fire, it was nicknamed "Old Droopy". A similar gun, possibly also salvaged from the Monongahela, is on display near the Bay View Club on the Naval Station. By 2006, only two elderly Cubans, Luis Delarosa and Harry Henry, still crossed the base's North East Gate daily to work on the base, because the Cuban government prohibited new recruitment since 1959. They both retired at the end of 2012. At the United Nations Human Rights Council in 2013, Cuba's Foreign Minister demanded the U.S. return the base. Current operations The military facilities at Guantanamo Bay have over 8,500 U.S. sailors and Marines stationed there. It is the only military base the U.S. maintains in a socialist country. In 2005, the U.S. Navy completed a $12 million wind-power project at the base, erecting four 950 kilowatt, wind turbines, reducing the need for diesel fuel to power the existing diesel generators (the base's primary electricity generation). In 2006, the wind turbines reduced diesel fuel consumption by annually. Units and commands Access to the Naval Station is very limited and must be pre-approved through the appropriate local chain of command with the Commander of the station as the final approval. Since berthing facilities are limited, visitors must be sponsored indicating that they have an approved residence for the duration of the visit. Resident units Headquarters, Naval Station Guantanamo Bay Customer Service Desk (CSD) Joint Task Force Guantanamo Headquarters, JTF Guantanamo Joint Detention Group Joint Intelligence Group Joint Medical Group U.S. Coast Guard Maritime Security Detachment Guantanamo Bay AFN Guantanamo Bay radio and television unit Marine Corps Security Force Company Naval Computer and Telecommunications Area Master Station Atlantic Detachment Guantanamo Bay Naval Hospital Guantanamo Bay Navy Supply Navy Security Forces SEABEE Detachment U.S. Coast Guard Aviation Detachment Guantanamo Bay Assigned units Fleet Composite Squadron Ten (VC-10) (1965–1993) U.S. Marine Corps Ground Defense Force (GDF) (1971–2000 [Redesignated as Marine Corps Security Forces Company on 1 Sep 2000]) Naval Security Group Activity (Company L) (1966–2001) Shore Intermediate Maintenance Activity (SIMA) (1903–1995) Fleet Training Group (FTG) (1943–1995) Homeported watercraft YC 1639 (open lighter) Leeward (YFB-92) (ferry boat) Windward (YFB-93) (ferry boat) YON 258 (non-self propelled fuel oil barge) (large harbor tug) LCU 1671 and MK-8: landing craft used as an alternate ferry for transportation to areas inaccessible by the primary ferry and for moving hazardous cargo. GTMO-5, GTMO-6 and GTMO-7 (50-ft. utility boats): used for personnel transportation during off-ferry hours. Civilian contractors Besides servicemembers, the base houses a large number of civilian contractors working for the military. Many of these contractors are migrant workers from Jamaica and the Philippines, and are thought to constitute up to 40% of the base's population. Major contractors working at NSGB have included the following: KBR Schuyler Line Navigation Company (SLNC) Satellite Communication Systems Incorporated Centerra EMCOR Islands Mechanical Contractor Munilla Construction Management RQ Construction MCM Construction J&J Worldwide Services Cargo shipping Ocean transportation is provided by Schuyler Line Navigation Company, a U.S. Flag Ocean Carrier. Schuyler Line operates under government contract to supply sustainment and building supplies to the base. Airfields There are two airfields within the base, Leeward Point Field and McCalla Field. Leeward Point Field is the active military airfield, with the ICAO code MUGM and IATA code NBW. McCalla Field was designated as the auxiliary landing field in 1970, but was no longer a viable airfield by the 1990s. Leeward Point Field Leeward Point Field was constructed in 1953 as part of Naval Air Station (NAS) Guantanamo Bay. Leeward Point Field has a single active runway, 10/28, measuring . The former runway, 9/27 was . Currently, Leeward Point Field operates several aircraft and helicopters supporting base operations. Leeward Point Field was home to Fleet Composite Squadron 10 (VC-10) until the unit was phased out in 1993. VC-10 was one of the last active-duty squadrons flying the Douglas A-4 Skyhawk. McCalla Field McCalla Field was established in 1931 and remained operational until 1970. The airfield was named for Bowman H. McCalla, who was a United States Navy admiral in charge of the Battle of Guantánamo Bay. The current field was expanded in 1941 when the original grass runway was replaced. Naval Air Station Guantanamo Bay was officially established 1 February 1941. Aircraft routinely operating out of McCalla included JRF-5, N3N, J2F, C-1 Trader, and dirigibles. The airfield was deactivated in the 1970s and was used to house Cuban and Haitian refugees beginning in the 1990s. Sometime between 1996 and 2001, the refugee camps were dismantled and the area became a collection of abandoned buildings. McCalla Field is now listed as a closed airfield. The USN aviation assets are now located across the bay at Leeward Point Field. After the events of September 11, the area was reevaluated as a possible location for a detention facility. Camp Justice is now located on the grounds of the former airfield. The area consists of 3 runways: 1/19 at , 14/32 at , and 10/28 at . The airport resides at an elevation of above mean sea level. Detention camp Beginning in the last quarter of the 20th century, the base was used to house Cuban and Haitian refugees intercepted on the high seas. In the early 1990s, it held refugees who fled Haiti after military forces overthrew president Jean-Bertrand Aristide. These refugees were held in a detainment area called Camp Bulkeley until United States district court Judge Sterling Johnson, Jr. declared the camp unconstitutional on 8 June 1993. This decision was later vacated. The last Haitian migrants departed Guantanamo on 1 November 1995. Beginning in 2002, following the U.S. invasion of Afghanistan, started in response to the September 11 attacks, a small portion of the base was used to detain several hundred individuals with ties or suspected ties to Islamic terrorist groups such as al-Qaeda and the Taliban. These detainees were not afforded POW status or facing formal charges but instead designated enemy combatants; assigned to confinement in Camp Delta, Camp Echo, Camp Iguana, and the now-closed Camp X-Ray. In litigation regarding the availability of fundamental rights to those imprisoned at the base, the U.S. Supreme Court has recognized that the detainees "have been imprisoned in territory over which the United States exercises exclusive jurisdiction and control." Therefore, the detainees have the fundamental right to due process of law under the Fifth Amendment. A district court has since held that the "Geneva Conventions applied to the Taliban detainees, but not to members of Al-Qaeda terrorist organization." On 10 June 2006, the Department of Defense reported that three Guantanamo Bay detainees took their own lives. The military reported the men hanged themselves with nooses made of sheets and clothes. A study published by Seton Hall Law's Center for Policy and Research, while making no conclusions regarding what actually transpired, asserts that the military investigation failed to address significant issues detailed in that report. On 6 September 2006, President George W. Bush announced that alleged or non-alleged combatants held by the CIA would be transferred to the custody of Department of Defense, and held at Guantanamo Prison. Of approximately 500 prisoners in Guantanamo Bay, only 10 have been tried by the Guantanamo military commission, but all cases have been stayed pending the adjustments being made to comply with the U.S. Supreme Court decision in Hamdi v. Rumsfeld. In January 2009, President Obama signed an executive order directing the closing of the Guantánamo detention camp within a year. This plan was thwarted for the time being on 20 May 2009, when the United States Senate voted to keep the prison at Guantanamo Bay open for the foreseeable future and forbid the transfer of any detainees to facilities in the United States. Senator Daniel Inouye, a Democrat from Hawaii and chairman of the appropriations committee, said he favored keeping Guantanamo open until Obama produced a "coherent plan for closing the prison." Consequently, Obama decided to postpone difficult decisions on the details for at least six months. On 7 March 2011, Obama issued a new executive order permitting indefinite detention of Guantánamo detainees. This decision was codified into federal law by provision added to the National Defense Authorization Act for Fiscal Year 2012. In early February 2021, the administration of U.S. President Joe Biden declared his intention to shut down the facility before he leaves office. In July 2021, an additional detainee was released. In June 2022, an Afghan prisoner held at Guantanamo Bay for about 15 years without trial was released. In February 2023, three prisoners were released from Guantanamo Bay. Majid Khan, who pled guilty before a Military Commission in 2012, had completed his sentence of 10 years, and was released to Belize. Brothers Abdul Al-Rahim Ghulam Rabbani and Mohammed Ahmad Ghulam Rabbani, whose detentions were deemed no longer necessary in order to protect the United States from a national security threat, were released and repatriated to the government of Pakistan. In March 2023, Ghassan Al Sharbi was released to his home country of Saudi Arabia from Guantanamo Bay, after being deemed no longer necessary to detain in order to protect the United States from a national security threat. Represented businesses Despite the prohibition on the establishment of "commercial or other enterprises" as stated in Article 3 of the second part of the lease, several recognized American food outlets have been opened at the military base. Most of the restaurants on the installation are franchises which are owned and operated by the Department of the Navy. Proceeds from these restaurants are used to support Morale, Welfare and Recreation (MWR) activities for service personnel and their families. These restaurants are located inside the base; as such, they are not accessible to Cubans. A Baskin-Robbins ice cream stand, which opened in the 1980s, was one of the first business franchises allowed on the base. In early 1986, the base added the first and only McDonald's restaurant within Cuba. A Hooters restaurant was opened in 1987. A Subway restaurant was opened in 2002. In 2004, a combined KFC & A&W restaurant was opened at the bowling alley and a Pizza Hut Express was added to the Windjammer Restaurant. There is also a cafe that sells Starbucks coffee, and there is a combined KFC & Taco Bell restaurant. Education and schooling Department of Defense Education Activity (DoDEA) provides for the education of dependent personnel with two schools. Both schools are named for Rear Admiral William Thomas Sampson. W.T. Sampson Elementary School serves grades K–5 and W. T. Sampson High School serves grades 6–12. The Villamar Child Development Center provides child care for dependents from six weeks to five years old. MWR operates a Youth Center that provides activities for dependents. Some former students of Guantánamo have shared stories of their experiences with the Guantánamo Public Memory Project. The 2013 documentary Guantanamo Circus directed by Christina Linhardt and Michael Rose reveals a glimpse of day-to-day life on GTMO as seen through the eyes of circus performers who visit the base. It is used as a reference by the Guantánamo Public Memory Project. Climate The bay has an annual rainfall of about . The amount of rainfall has resulted in the base being classified as a semi-arid desert environment. The annual average high temperature on the base is , the annual average low is . Notable people Notable people born at the naval base include actor Peter Bergman and American guitarist Isaac Guillory. See also COVID-19 pandemic in the Guantanamo Bay Naval Base Cuba–United States relations Platt Amendment A Few Good Men – A play by Aaron Sorkin, and later a film adaptation about an unlawful hazing incident taking place amongst marines at the naval base in 1986. The Road to Guantanamo – A docudrama directed by Michael Winterbottom about the incarceration of three British detainees at the naval base. Cuban–American Treaty of Relations Panama Canal Zone References Further reading Jonathan M. Hansen, Guantánamo: An American History. New York: Hill and Wang, 2011. Alfred de Zayas, "The Status of Guantanamo Bay and the Status of the Detainees" in University of British Columbia Law Review, vol. 37, July 2004, pp. 277–34; A de Zayas Guantanamo Naval Base in Max Planck Encyclopedia of Public International Law, Oxford University Press 2012) Mason, Theodore K. Beyond the Cactus Curtain: The Story of Guantanamo Bay. New York: Dodd, Mead, 1984. External links Virtual 3D Walkthrough of Camp Delta (from the Art project Zone*Interdite) Guantanamo Bay 1898 establishments in Cuba 1898 establishments in the United States Banana Wars Buildings and structures in Guantánamo Province Geography of Guantánamo Province International territorial disputes of the United States Military installations established in 1898 Military installations of the United States in Cuba Spanish–American War Territorial disputes of Cuba United States Navy installations War on terror
13212
https://en.wikipedia.org/wiki/History%20of%20Europe
History of Europe
The history of Europe is traditionally divided into four time periods: prehistoric Europe (prior to about 800 BC), classical antiquity (800 BC to AD 500), the Middle Ages (AD 500 to AD 1500), and the modern era (since AD 1500). The first early European modern humans appear in the fossil record about 48,000 years ago, during the Paleolithic Era. People from this period left behind numerous artifacts, including works of art, burial sites, and tools, allowing some reconstruction of their society. Settled agriculture marked the Neolithic Era, which spread slowly across Europe from southeast to the north and west. The later Neolithic period saw the introduction of early metallurgy and the use of copper-based tools and weapons, and the building of megalithic structures, as exemplified by Stonehenge. During the Indo-European migrations, Europe saw migrations from the east and southeast. The period known as classical antiquity began with the emergence of the city-states of ancient Greece. Some of the earliest examples of literature, history, and philosophy come from the writings of the ancient Greeks, such as Homer, Herodotus, and Plato. Later, the Roman Empire came to dominate the entire Mediterranean basin. The Migration Period of the Germanic people began in the late 4th century AD and made gradual incursions into various parts of the Roman Empire. As these migratory people settled down and formed state societies of their own, this marked the transition period out of the classical era. The Fall of the Western Roman Empire in AD 476 traditionally marks the start of the Middle Ages. While the Eastern Roman Empire would continue for another 1000 years, the former lands of the Western Empire would be fragmented into a number of different states. At the same time, the early Slavs began to become established as a distinct group in the central and eastern parts of Europe. The first great empire of the Middle Ages was the Frankish Empire of Charlemagne, while the Islamic conquest of Iberia established Al-Andalus. The Viking Age saw a second great migration of Norse peoples. Attempts to retake the Levant from the Muslim states that occupied it made the High Middle Ages the age of The Crusades, while the political system of feudalism came to its height. The Late Middle Ages were marked by large population declines, as Europe was threatened by the Bubonic Plague, as well as invasions by the Mongol peoples from the Eurasian Steppe. At the end of the Middle Ages, there was a transitional period, known as the Renaissance. Early Modern Europe is usually dated to the end of the 15th century. Technological changes such as gunpowder and the printing press changed how warfare was conducted and how knowledge was preserved and disseminated. The Protestant Reformation saw the fragmentation of religious thought, leading to religious wars. The Age of Exploration led to colonization, and the exploitation of the people and resources of colonies brought resources and wealth to Europe. After 1800, the Industrial Revolution brought capital accumulation and rapid urbanization to Western Europe, while several countries transitioned away from absolutist rule to parliamentary regimes. The Age of Revolutions saw long-established political systems upset and turned over. In the 20th century, World War I led to a remaking of the map of Europe as the large Empires were broken up into nation-states. Lingering political issues would lead to World War II, during which Nazi Germany perpetrated the Holocaust. After World War II, during the Cold War, most of Europe became divided by the Iron Curtain in two military blocs: NATO and the Warsaw Pact. The post-war period saw decolonization as Western European colonial empires were dismantled. The post-war period also featured the gradual development of the European integration process, which led to the creation of the European Union; this extended to Eastern European countries after the Fall of the Berlin Wall. The 21st century saw the European debt crisis, the withdrawal of the United Kingdom from the European Union and the Russian invasion of Ukraine. Overview During the Neolithic era (starting at c. 7000 BC) and the time of the Indo-European migrations (starting at c. 4000 BC.) Europe saw massive migrations from the east and southeast which also brought agriculture, new technologies, and the Indo-European languages, primarily through the areas of the Balkan peninsula and the Black sea region. Some of the most prominent civilizations of the late prehistoric Europe were the Minoan and the Mycenaean, which flourished during the Bronze Age until they collapsed in a short period of time around 1200 BC. The period known as classical antiquity began with the emergence of the city-states of Ancient Greece. After ultimately checking the Persian advance in Europe through the Greco-Persian Wars in the 5th century BC, Greek influence reached its zenith under the expansive empire of Alexander the Great, spreading throughout Asia, Africa, and other parts of Europe. The Thracians, their powerful Odrysian kingdom, distinct culture and architecture were long present in Southeast Europe. The Roman Empire came to dominate the entire Mediterranean basin. By 300 AD the Roman Empire was divided into the Western and Eastern empires. During the 4th and 5th centuries, the Germanic peoples of Northern Europe, pressed by the Huns, grew in strength and led repeated attacks that resulted in the Fall of the Western Roman Empire. The Western empire's collapse in AD 476 traditionally marks the end of the classical period and the start of the Middle Ages. In Western Europe, Germanic peoples became more powerful in the remnants of the former Western Roman Empire and established kingdoms and empires of their own. Of all of the Germanic peoples, the Franks would rise to a position of hegemony over Western Europe, the Frankish Empire reaching its peak under Charlemagne around 800. This empire was later divided into several parts; West Francia would evolve into the Kingdom of France, while East Francia would evolve into the Holy Roman Empire, a precursor to modern Germany and Italy. The British Isles were the site of several large-scale migrations. The Byzantine Empire – the eastern part of the Roman Empire, with its capital Constantinople, flourished for the next 1000 years. During most of its existence, the empire was the most dominant of all, and also the most powerful economic, cultural, and military force in Europe. The powerful and long lived Bulgarian Empire was its main competitor in the region of Southeast Europe for centuries. Byzantine art, architecture, political dominance, and Bulgarian cultural and linguistic achievements left a great legacy in Orthodox and Slavic Europe and beyond through the Middle Ages to this day. The Viking Age, a period of migrations of Scandinavian peoples, occurred from the late 8th century to the middle 11th century. The Normans, descendants of the Vikings who settled in Northern France, had a significant impact on many parts of Europe, from the Norman conquest of England to Sicily. The Rus' people founded Kievan Rus', which evolved into Russia. After 1000 the Crusades were a series of religiously motivated military expeditions originally intended to bring the Levant back under Christian rule. The Crusaders opened trade routes which enabled the merchant republics of Genoa and Venice to become major economic powers. The Reconquista, a related movement, worked to reconquer Iberia for Christendom. Eastern Europe in the High Middle Ages was dominated by the rise and fall of the Mongol Empire. Led by Genghis Khan, the Mongols were a group of steppe nomads who established a decentralized empire which, at its height, extended from China in the east to the Black and Baltic Seas in Europe. As Mongol power waned towards the Late Middle Ages, the Grand Duchy of Moscow rose to become the strongest of the numerous Russian principalities and republics and would grow into the Tsardom of Russia in 1547. The Late Middle Ages represented a period of upheaval in Europe. The pandemic known as the Black Death and an associated famine caused demographic catastrophe in Europe as the population plummeted. Dynastic struggles and wars of conquest kept many of the states of Europe at war for much of the period. In Scandinavia, the Kalmar Union dominated the political landscape, while England fought with Scotland in the Wars of Scottish Independence and with France in the Hundred Years' War. In Central Europe, the Polish–Lithuanian Commonwealth became a large territorial empire, while the Holy Roman Empire, which was an elective monarchy, came to be dominated for centuries by the House of Habsburg. Russia continued to expand southward and eastward into former Mongol lands. In the Balkans, the Ottoman Empire overran Byzantine lands, culminating in the Fall of Constantinople in 1453, which historians mark as the end of the Middle Ages. Beginning in the 14th century in Florence and later spreading through Europe, a Renaissance of knowledge challenged traditional doctrines in science and theology. The rediscovery of classical Greek and Roman knowledge had an enormous liberating effect on intellectuals. Simultaneously, the Protestant Reformation under German Martin Luther questioned Papal authority. Henry VIII seized control of the English Church and its lands. The European religious wars were fought between German and Spanish rulers. The Reconquista ended Muslim rule in Iberia. By the 1490s a series of oceanic explorations marked the Age of Discovery, establishing direct links with Africa, the Americas, and Asia. Religious wars continued to be fought in Europe, until the 1648 Peace of Westphalia. The Spanish crown maintained its hegemony in Europe and was the leading power on the continent until the signing of the Treaty of the Pyrenees, which ended a conflict between Spain and France that had begun during the Thirty Years' War. An unprecedented series of major wars and political revolutions took place around Europe and the world in the period between 1610 and 1700. The Industrial Revolution began in Britain, based on coal, steam, and textile mills. Political change in continental Europe was spurred by the French Revolution under the motto liberté, égalité, fraternité. Napoleon Bonaparte took control, made many reforms inside France, and transformed Western Europe. But his rise stimulated both nationalism and reaction and he was defeated in 1814–15 as the old royal conservatives returned to power. The period between 1815 and 1871 saw revolutionary attempts in much of Europe (apart from Britain). They all failed however. As industrial work forces grew in Western Europe, socialism and trade union activity developed. The last vestiges of serfdom were abolished in Russia in 1861. Greece and the other Balkan nations began a long slow road to independence from the Ottoman Empire, starting in the 1820s. Italy was unified in its Risorgimento in 1860. After the Franco-Prussian War of 1870–71, Otto von Bismarck unified the German states into an empire that was politically and militarily dominant until 1914. Most of Europe scrambled for imperial colonies in Africa and Asia in the Age of Empire. Britain and France built the largest empires, while diplomats ensured there were no major wars in Europe, apart from the Crimean War of the 1850s. The outbreak of the First World War in 1914 was precipitated by the rise of nationalism in Southeastern Europe as the Great Powers took sides. The 1917 October Revolution led the Russian Empire to become the world's first communist state, the Soviet Union. The Allies, led by Britain, France, and the United States, defeated the Central Powers, led by the German Empire and Austria-Hungary, in 1918. During the Paris Peace Conference the Big Four imposed their terms in a series of treaties, especially the Treaty of Versailles. The war's human and material devastation was unprecedented. Germany lost its overseas empire and several provinces, had to pay large reparations, and was humiliated by the victors. They in turn had large debts to the United States. The 1920s were prosperous until 1929 when the Great Depression broke out, which led to the collapse of democracy in many European states. The Nazi regime under Adolf Hitler came to power in 1933, rearmed Germany, and along with Mussolini's Italy sought to assert themselves on the continent. Other nations, who had not taken to the attractions of fascism, sought to avoid conflict. They set boundaries of appeasement, which Hitler continually ignored. The Second World War began. The war ended with the defeat of the Axis powers but the threat of more conflict was recognised before the war's end. Many from the US were suspicious of how the USSR would treat the peace – in the USSR there was paranoia at US forces in Europe. Eastern Front/Western Front meetings among leaders in Yalta proved inconclusive. In the closing months of the war there was a race to the finish. The territories liberated from the Nazis by troops from the USSR found they had exchanged fascism for socialism. The USSR, however, would not leave those territories for forty years. The USSR claimed they needed buffer states between them and the nascent NATO. In the west, the term Iron Curtain entered the language. The United States launched the Marshall Plan from 1948 to 1951 and NATO from 1949, and rebuilt industrial economies that all were thriving by the 1950s. France and West Germany took the lead in forming the European Economic Community, which eventually became the European Union (EU). Secularization saw the weakening of Protestant and Catholic churches across most of Europe, except where they were symbols of reaction, as in Poland. The Revolutions of 1989 brought an end to both Soviet hegemony and socialism in Eastern Europe, the resulting capitalist restoration engendering economic and social devastation for the people. Germany was reunited, Europe's integration deepened, and both NATO and the EU expanded to the east. The EU came under increasing pressure because of the worldwide recession after 2008. In 2022, Russia would invade Ukraine. Prehistory of Europe Paleolithic Homo erectus migrated from Africa to Europe before the emergence of modern humans. Homo erectus georgicus, which lived roughly 1.8 million years ago in Georgia, is the earliest hominid to have been discovered in Europe. Lézignan-la-Cèbe in France, Orce in Spain, Monte Poggiolo in Italy and Kozarnika in Bulgaria are among the oldest Palaeolithic sites in Europe. The earliest appearance of anatomically modern people in Europe has been dated to 45,000 BC, referred to as the Early European modern humans. The earliest sites in Europe are Riparo Mochi (Italy), Swabian Jura (Germany), and Isturitz (France). Some locally developed transitional cultures (Uluzzian in Italy and Greece, Altmühlian in Germany, Szeletian in Central Europe and Châtelperronian in the southwest) use clearly Upper Palaeolithic technologies at very early dates. Nevertheless, the definitive advance of these technologies is made by the Aurignacian culture. The origins of this culture can be located in the Levant (Ahmarian) and Hungary (first full Aurignacian). By 35,000 BC, the Aurignacian culture and its technology had extended through most of Europe. The last Neanderthals seem to have been forced to retreat during this process to the southern half of the Iberian Peninsula. Around 29,000 BC a new technology/culture appeared in the western region of Europe: the Gravettian. This technology/culture has been theorised to have come with migrations of people from the Balkans (see Kozarnika). Around 16,000 BC, Europe witnessed the appearance of a new culture, known as Magdalenian, possibly rooted in the old Gravettian. This culture soon superseded the Solutrean area and the Gravettian of mainly France, Spain, Germany, Italy, Poland, Portugal and Ukraine. The Hamburg culture prevailed in Northern Europe in the 14th and the 13th millennium BC as the Creswellian (also termed the British Late Magdalenian) did shortly after in the British Isles. Around 12,500 BC, the Würm glaciation ended. Slowly, through the following millennia, temperatures and sea levels rose, changing the environment of prehistoric people. Nevertheless, Magdalenian culture persisted until c. 10,000 BC, when it quickly evolved into two microlithist cultures: Azilian (Federmesser), in Spain and southern France, and then Sauveterrian, in southern France and Tardenoisian in Central Europe, while in Northern Europe the Lyngby complex succeeded the Hamburg culture with the influence of the Federmesser group as well. Neolithic and Copper Age Evidence of permanent settlement dates from the 8th millennium BC in the Balkans. The Neolithic reached Central Europe in the 6th millennium BC and parts of Northern Europe in the 5th and 4th millenniums BC. The modern indigenous populations of Europe are largely descended from three distinct lineages: Mesolithic hunter-gatherers, a derivative of the Cro-Magnon population of Europe, Early European Farmers who migrated from Anatolia during the Neolithic Revolution, and Yamnaya pastoralists who expanded into Europe in the context of the Indo-European expansion. The genetic makeup of speakers of the Uralic language family in northern Europe was shaped by migration from Siberia that began at least 3,500 years ago. The Indo-European migrations started in Southeast Europe at around c. 4200 BC. through the areas around the Black sea and the Balkan peninsula. In the next 3000 years the Indo-European languages expanded through Europe. Around this time, in the 5th millennium BC the Varna culture evolved. In 4700 – 4200 BC, the Solnitsata town, believed to be the oldest prehistoric town in Europe, flourished. In the Varna Necropolis – a burial site from 4569 to 4340 BC and one of the most important archaeological sites in world prehistory, was found the oldest gold treasure (elaborated golden objects) in the world. Recently discovered golden artifacts in another site in Bulgaria near Durankulak appear to be 7,000 years old. Several prehistoric Bulgarian finds are considered no less old – the golden treasures of Hotnitsa, artifacts from the Kurgan settlement of Yunatsite near Pazardzhik, the golden treasure Sakar, as well as beads and gold jewelry found in the Kurgan settlement of Provadia – Solnitsata (“salt pit”). However, Varna gold is most often called the oldest since this treasure is the largest and most diverse. Ancient Europe Bronze Age The first well-known literate civilization in Europe was that of the Minoans. The Minoan civilization was a Bronze Age civilization that arose on the island of Crete and flourished from approximately the 27th century BC to the 15th century BC. It was rediscovered at the beginning of the 20th century through the work of the British archaeologist Arthur Evans. Will Durant referred to it as "the first link in the European chain". The Minoans were replaced by the Mycenaean civilization which flourished during the period roughly between 1600 BC, when Helladic culture in mainland Greece was transformed under influences from Minoan Crete, and 1100 BC. The major Mycenaean cities were Mycenae and Tiryns in Argolis, Pylos in Messenia, Athens in Attica, Thebes and Orchomenus in Boeotia, and Iolkos in Thessaly. In Crete, the Mycenaeans occupied Knossos. Mycenaean settlement sites also appeared in Epirus, Macedonia, on islands in the Aegean Sea, on the coast of Asia Minor, the Levant, Cyprus and Italy. Mycenaean artefacts have been found well outside the limits of the Mycenean world. Quite unlike the Minoans, whose society benefited from trade, the Mycenaeans advanced through conquest. Mycenaean civilization was dominated by a warrior aristocracy. Around 1400 BC, the Mycenaeans extended their control to Crete, the centre of the Minoan civilization, and adopted a form of the Minoan script (called Linear A) to write their early form of Greek in Linear B. The Mycenaean civilization perished with the collapse of Bronze-Age civilization on the eastern shores of the Mediterranean Sea. The collapse is commonly attributed to the Dorian invasion, although other theories describing natural disasters and climate change have been advanced as well. Whatever the causes, the Mycenaean civilization had definitely disappeared after LH III C, when the sites of Mycenae and Tiryns were again destroyed and lost their importance. This end, during the last years of the 12th century BC, occurred after a slow decline of the Mycenaean civilization, which lasted many years before dying out. The beginning of the 11th century BC opened a new context, that of the protogeometric, the beginning of the geometric period, the Greek Dark Ages of traditional historiography. The Bronze Age collapse may be seen in the context of a technological history that saw the slow spread of ironworking technology from present-day Bulgaria and Romania in the 13th and the 12th centuries BC. Classical Antiquity The Greeks and the Romans left a legacy in Europe which is evident in European languages, thought, visual arts and law. Ancient Greece was a collection of city-states, out of which the original form of democracy developed. Athens was the most powerful and developed city, and a cradle of learning from the time of Pericles. Citizens' forums debated and legislated policy of the state, and from here arose some of the most notable classical philosophers, such as Socrates, Plato, and Aristotle, the last of whom taught Alexander the Great. Through his military campaigns, the king of the kingdom of Macedon, Alexander, spread Hellenistic culture and learning to the banks of the River Indus. Meanwhile, the Roman Republic strengthened through victory over Carthage in the Punic Wars. Greek wisdom passed into Roman institutions, as Athens itself was absorbed under the banner of the Senate and People of Rome (SPQR). The Romans expanded their domains from Anatolia in the east to Britannia in the west. In 44 BC as it approached its height, its dictator Julius Caesar was murdered by senators in an attempt to restore the Republic. In the ensuing turmoil, Octavian (ruled as Augustus; and as divi filius, or Son of God, as Julius had adopted him as an heir) usurped the reins of power and fought the Roman Senate. While proclaiming the rebirth of the Republic, he had ushered in the transfer of the Roman state from a republic to an empire, the Roman Empire, which lasted for nearly 15 centuries until the fall of the Eastern Roman Empire. Ancient Greece The Hellenic civilisation was a collection of city-states or poleis with different governments and cultures that achieved notable developments in government, philosophy, science, mathematics, politics, sports, theatre and music. The most powerful city-states were Athens, Sparta, Thebes, Corinth, and Syracuse. Athens was a powerful Hellenic city-state and governed itself with an early form of direct democracy invented by Cleisthenes; the citizens of Athens voted on legislation and executive bills themselves. Athens was the home of Socrates, Plato, and the Platonic Academy. The Hellenic city-states established colonies on the shores of the Black Sea and the Mediterranean Sea (Asian Minor, Sicily, and Southern Italy in Magna Graecia). By the late 6th century BC, all the Greek city states in Asia Minor had been incorporated into the Persian Empire, while the latter had made territorial gains in the Balkans (such as Macedon, Thrace, Paeonia, etc.) and Eastern Europe proper as well. In the course of the 5th century BC, some of the Greek city states attempted to overthrow Persian rule in the Ionian Revolt, which failed. This sparked the first Persian invasion of mainland Greece. At some point during the ensuing Greco-Persian Wars, namely during the Second Persian invasion of Greece, and precisely after the Battle of Thermopylae and the Battle of Artemisium, almost all of Greece to the north of the Isthmus of Corinth had been overrun by the Persians, but the Greek city states reached a decisive victory at the Battle of Plataea. With the end of the Greco-Persian wars, the Persians were eventually decisively forced to withdraw from their territories in Europe. The Greco-Persian Wars and the victory of the Greek city states directly influenced the entire further course of European history and would set its further tone. Some Greek city-states formed the Delian League to continue fighting Persia, but Athens' position as leader of this league led Sparta to form the rival Peloponnesian League. The Peloponnesian Wars ensued, and the Peloponnesian League was victorious. Subsequently, discontent with Spartan hegemony led to the Corinthian War and the defeat of Sparta at the Battle of Leuctra. At the same time at the north ruled the Thracian Odrysian Kingdom between the 5th century BC and the 1st century AD. Hellenic infighting left Greek city states vulnerable, and Philip II of Macedon united the Greek city states under his control. The son of Philip II, known as Alexander the Great, invaded neighboring Persia, toppled and incorporated its domains, as well as invading Egypt and going as far off as India, increasing contact with people and cultures in these regions that marked the beginning of the Hellenistic period. After the death of Alexander the Great, his empire split into multiple kingdoms ruled by his generals, the Diadochi. The Diadochi fought against each other in a series of conflicts called the Wars of the Diadochi. In the beginning of the 2nd century BC, only three major kingdoms remained: the Ptolemaic Egypt, the Seleucid Empire and Macedonia. These kingdoms spread Greek culture to regions as far away as Bactria. Ancient Rome The rise of Rome Much of Greek learning was assimilated by the nascent Roman state as it expanded outward from Italy, taking advantage of its enemies' inability to unite: the only challenge to Roman ascent came from the Phoenician colony of Carthage, and its defeats in the three Punic Wars marked the start of Roman hegemony. First governed by kings, then as a senatorial republic (the Roman Republic), Rome finally became an empire at the end of the 1st century BC, under Augustus and his authoritarian successors. The Roman Empire had its centre in the Mediterranean, controlling all the countries on its shores; the northern border was marked by the Rhine and Danube rivers. Under the emperor Trajan (2nd century AD) the empire reached its maximum expansion, controlling approximately of land surface, including Italia, Gallia, Dalmatia, Aquitania, Britannia, Baetica, Hispania, Thrace, Macedonia, Greece, Moesia, Dacia, Pannonia, Egypt, Asia Minor, Cappadocia, Armenia, Caucasus, North Africa, Levant and parts of Mesopotamia. Pax Romana, a period of peace, civilisation and an efficient centralised government in the subject territories ended in the 3rd century, when a series of civil wars undermined Rome's economic and social strength. In the 4th century, the emperors Diocletian and Constantine were able to slow down the process of decline by splitting the empire into a Western part with a capital in Rome and an Eastern part with the capital in Byzantium, or Constantinople (now Istanbul). Constantinople is generally considered to be the center of "Eastern Orthodox civilization". Whereas Diocletian severely persecuted Christianity, Constantine declared an official end to state-sponsored persecution of Christians in 313 with the Edict of Milan, thus setting the stage for the Church to become the state church of the Roman Empire in about 380. Decline of the Roman Empire The Roman Empire had been repeatedly attacked by invading armies from Northern Europe and in 476, Rome finally fell. Romulus Augustus, the last emperor of the Western Roman Empire, surrendered to the Germanic King Odoacer. The British historian Edward Gibbon argued in The History of the Decline and Fall of the Roman Empire (1776) that the Romans had become decadent and had lost civic virtue. Gibbon said that the adoption of Christianity meant belief in a better life after death, and therefore made people lazy and indifferent to the present. "From the eighteenth century onward", Glen W. Bowersock has remarked, "we have been obsessed with the fall: it has been valued as an archetype for every perceived decline, and, hence, as a symbol for our own fears." It remains one of the greatest historical questions, and has a tradition rich in scholarly interest. Some other notable dates are the Battle of Adrianople in 378, the death of Theodosius I in 395 (the last time the Roman Empire was politically unified), the crossing of the Rhine in 406 by Germanic tribes after the withdrawal of the legions to defend Italy against Alaric I, the death of Stilicho in 408, followed by the disintegration of the western legions, the death of Justinian I, the last Roman emperor who tried to reconquer the west, in 565, and the coming of Islam after 632. Many scholars maintain that rather than a "fall", the changes can more accurately be described as a complex transformation. Over time many theories have been proposed on why the Empire fell, or whether indeed it fell at all. Late Antiquity and Migration Period When Emperor Constantine had reconquered Rome under the banner of the cross in 312, he soon afterwards issued the Edict of Milan in 313 (preceded by the Edict of Serdica in 311), declaring the legality of Christianity in the Roman Empire. In addition, Constantine officially shifted the capital of the Roman Empire from Rome to the Greek town of Byzantium, which he renamed Nova Roma – it was later named Constantinople ("City of Constantine"). Theodosius I, who had made Christianity the official religion of the Roman Empire, would be the last emperor to preside over a united Roman Empire, until his death in 395. The empire was split into two-halves: the Western Roman Empire centred in Ravenna, and the Eastern Roman Empire (later to be referred to as the Byzantine Empire) centred in Constantinople. The Roman Empire was repeatedly attacked by Hunnic, Germanic, Slavic and other "barbarian" tribes (see: Migration Period), and in 476 finally the Western part fell to the Heruli chieftain Odoacer. Roman authority in the Western part of the empire had collapsed, and a power vacuum left in the wake of this collapse; the central organization, institutions, laws and power of Rome had broken down, resulting in many areas being open to invasion by migrating tribes. Over time, feudalism and manorialism arose, two interlocking institutions that provided for division of land and labour, as well as a broad if uneven hierarchy of law and protection. These localised hierarchies were based on the bond of common people to the land on which they worked, and to a lord, who would provide and administer both local law to settle disputes among the peasants, as well as protection from outside invaders. Unlike under Roman rule, with its standard laws and military across the empire and its great bureaucracy to administer them and collect taxes, each lord (although having obligations to a higher lord) was largely sovereign in his domain. A peasant's lot could vary greatly depending on the leadership skills and attitudes to justice of the lord toward his people. Tithes or rents were paid to the lord, who in turn owed resources, and armed men in times of war, to his lord, perhaps a regional prince. However, the levels of hierarchy were varied over time and place. The western provinces soon were to be dominated by three great powers: first, the Franks (Merovingian dynasty) in Francia 481–843 AD, which covered much of present France and Germany; second, the Visigothic kingdom 418–711 AD in the Iberian Peninsula (modern Spain); and third, the Ostrogothic kingdom 493–553 AD in Italy and parts of the western Balkans. The Ostrogoths were later replaced by the Kingdom of the Lombards 568–774 AD. These new powers of the west built upon the Roman traditions until they evolved into a synthesis of Roman and Germanic cultures. Although these powers covered large territories, they did not have the great resources and bureaucracy of the Roman empire to control regions and localities. The ongoing invasions and boundary disputes usually meant a more risky and varying life than that under the empire. This meant that in general more power and responsibilities were left to local lords. On the other hand, it also meant more freedom, particularly in more remote areas. In Italy, Theodoric the Great began the cultural romanisation of the new world he had constructed. He made Ravenna a centre of Romano-Greek culture of art and his court fostered a flowering of literature and philosophy in Latin. In Iberia, King Chindasuinth created the Visigothic Code. In the Eastern part the dominant state was the remaining Eastern Roman Empire. In the feudal system, new princes and kings arose, the most powerful of which was arguably the Frankish ruler Charlemagne. In 800, Charlemagne, reinforced by his massive territorial conquests, was crowned Emperor of the Romans (Imperator Romanorum) by Pope Leo III, effectively solidifying his power in western Europe. Charlemagne's reign marked the beginning of a new Germanic Roman Empire in the west, the Holy Roman Empire. Outside his borders, new forces were gathering. The Kievan Rus' were marking out their territory, a Great Moravia was growing, while the Angles and the Saxons were securing their borders. For the duration of the 6th century, the Eastern Roman Empire was embroiled in a series of deadly conflicts, first with the Persian Sassanid Empire (see Roman–Persian Wars), followed by the onslaught of the arising Islamic Caliphate (Rashidun and Umayyad). By 650, the provinces of Egypt, Palestine and Syria were lost to the Muslim forces, followed by Hispania and southern Italy in the 7th and 8th centuries (see Muslim conquests). The Arab invasion from the east was stopped after the intervention of the Bulgarian Empire (see Han Tervel). Post-classical Europe The Middle Ages are commonly dated from the fall of the Western Roman Empire (or by some scholars, before that) in the 5th century to the beginning of the early modern period in the 16th century marked by the rise of nation states, the division of Western Christianity in the Reformation, the rise of humanism in the Italian Renaissance, and the beginnings of European overseas expansion which allowed for the Columbian Exchange. Byzantium Many consider Emperor Constantine I (reigned 306–337) to be the first "Byzantine emperor". It was he who moved the imperial capital in 324 from Nicomedia to Byzantium, which re-founded as Constantinople, or Nova Roma ("New Rome"). The city of Rome itself had not served as the capital since the reign of Diocletian (284–305). Some date the beginnings of the Empire to the reign of Theodosius I (379–395) and Christianity's official supplanting of the pagan Roman religion, or following his death in 395, when the empire was split into two parts, with capitals in Rome and Constantinople. Others place it yet later in 476, when Romulus Augustulus, traditionally considered the last western emperor, was deposed, thus leaving sole imperial authority with the emperor in the Greek East. Others point to the reorganisation of the empire in the time of Heraclius (c. 620) when Latin titles and usages were officially replaced with Greek versions. In any case, the changeover was gradual and by 330, when Constantine inaugurated his new capital, the process of hellenization and increasing Christianisation was already under way. The Empire is generally considered to have ended after the fall of Constantinople to the Ottoman Turks in 1453. The Plague of Justinian was a pandemic that afflicted the Byzantine Empire, including its capital Constantinople, in the years 541–542. It is estimated that the Plague of Justinian killed as many as 100 million people across the world. It caused Europe's population to drop by around 50% between 541 and 700. It also may have contributed to the success of the Muslim conquests. During most of its existence, the Byzantine Empire was one of the most powerful economic, cultural, and military forces in Europe, and Constantinople was one of the largest and wealthiest cities in Europe. Early Middle Ages The Early Middle Ages span roughly five centuries from 500 to 1000. In the East and Southeast of Europe new dominant states formed: the Avar Khaganate (567–after 822), Old Great Bulgaria (632–668), the Khazar Khaganate (c. 650–969) and Danube Bulgaria (founded by Asparuh in 680) were constantly rivaling the hegemony of the Byzantine Empire. From the 7th century Byzantine history was greatly affected by the rise of Islam and the Caliphates. Muslim Arabs first invaded historically Roman territory under Abū Bakr, first Caliph of the Rashidun Caliphate, who entered Roman Syria and Roman Mesopotamia. As the Byzantines and neighboring Sasanids were severely weakened by the time, amongst the most important reason(s) being the protracted, centuries-lasting and frequent Byzantine–Sasanian wars, which included the climactic Byzantine–Sasanian War of 602–628, under Umar, the second Caliph, the Muslims entirely toppled the Sasanid Persian Empire, and decisively conquered Syria and Mesopotamia, as well as Roman Palestine, Roman Egypt, and parts of Asia Minor and Roman North Africa. In the mid 7th century AD, following the Muslim conquest of Persia, Islam penetrated into the Caucasus region, of which parts would later permanently become part of Russia. This trend, which included the conquests by the invading Muslim forces and by that the spread of Islam as well continued under Umar's successors and under the Umayyad Caliphate, which conquered the rest of Mediterranean North Africa and most of the Iberian Peninsula. Over the next centuries Muslim forces were able to take further European territory, including Cyprus, Malta, Crete, and Sicily and parts of southern Italy. The Muslim conquest of Hispania began when the Moors (Berbers and Arabs) invaded the Christian Visigothic kingdom of Hispania in the year 711, under the Berber general Tariq ibn Ziyad. They landed at Gibraltar on 30 April and worked their way northward. Tariq's forces were joined the next year by those of his Arab superior, Musa ibn Nusair. During the eight-year campaign most of the Iberian Peninsula was brought under Muslim rule – save for small areas in the northwest (Asturias) and largely Basque regions in the Pyrenees. In 711, Visigothic Hispania was very weakened because it was immersed in a serious internal crisis caused by a war of succession to the throne involving two Visigoth suitors. The Muslims took advantage of the crisis within the Hispano-Visigothic society to carry out their conquests. This territory, under the Arab name Al-Andalus, became part of the expanding Umayyad empire. The second siege of Constantinople (717) ended unsuccessfully after the intervention of Tervel of Bulgaria and weakened the Umayyad dynasty and reduced their prestige. In 722 Don Pelayo, a nobleman of Visigothic origin, formed an army of 300 Astur soldiers, to confront Munuza's Muslim troops. In the battle of Covadonga, the Astures defeated the Arab-Moors, who decided to retire. The Christian victory marked the beginning of the Reconquista and the establishment of the Kingdom of Asturias, whose first sovereign was Don Pelayo. The conquerors intended to continue their expansion in Europe and move northeast across the Pyrenees, but were defeated by the Frankish leader Charles Martel at the Battle of Poitiers in 732. The Umayyads were overthrown in 750 by the 'Abbāsids, and, in 756, the Umayyads established an independent emirate in the Iberian Peninsula. Feudal Christendom The Holy Roman Empire emerged around 800, as Charlemagne, King of the Franks and part of the Carolingian dynasty, was crowned by the pope as emperor. His empire based in modern France, the Low Countries and Germany expanded into modern Hungary, Italy, Bohemia, Lower Saxony and Spain. He and his father received substantial help from an alliance with the Pope, who wanted help against the Lombards. His death marked the beginning of the end of the dynasty, which collapsed entirely by 888. The fragmentation of power led to semi-autonomy in the region, and has been defined as a critical starting point for the formation of states in Europe. To the east, Bulgaria was established in 681 and became the first Slavic country. The powerful Bulgarian Empire was the main rival of Byzantium for control of the Balkans for centuries and from the 9th century became the cultural centre of Slavic Europe. The Empire created the Cyrillic script during the 9th century AD, at the Preslav Literary School, and experienced the Golden Age of Bulgarian cultural prosperity during the reign of emperor Simeon I the Great (893–927). Two states, Great Moravia and Kievan Rus', emerged among the Slavic peoples respectively in the 9th century. In the late 9th and 10th centuries, northern and western Europe felt the burgeoning power and influence of the Vikings who raided, traded, conquered and settled swiftly and efficiently with their advanced seagoing vessels such as the longships. The Vikings had left a cultural influence on the Anglo-Saxons and Franks as well as the Scots. The Hungarians pillaged mainland Europe, the Pechenegs raided Bulgaria, Rus States and the Arab states. In the 10th century independent kingdoms were established in Central Europe including Poland and the newly settled Kingdom of Hungary. The Kingdom of Croatia also appeared in the Balkans. The subsequent period, ending around 1000, saw the further growth of feudalism, which weakened the Holy Roman Empire. In eastern Europe, Volga Bulgaria became an Islamic state in 921, after Almış I converted to Islam under the missionary efforts of Ahmad ibn Fadlan. Slavery in the early medieval period had mostly died out in western Europe by about the year 1000 AD, replaced by serfdom. It lingered longer in England and in peripheral areas linked to the Muslim world, where slavery continued to flourish. Church rules suppressed slavery of Christians. Most historians argue the transition was quite abrupt around 1000, but some see a gradual transition from about 300 to 1000. High Middle Ages The slumber of the Dark Ages was shaken by a renewed crisis in the Church. In 1054, the East–West Schism, an insoluble split, occurred between the two remaining Christian seats in Rome and Constantinople (modern Istanbul). The High Middle Ages of the 11th, 12th, and 13th centuries show a rapidly increasing population of Europe, which caused great social and political change from the preceding era. By 1250, the robust population increase greatly benefited the economy, reaching levels it would not see again in some areas until the 19th century. From about the year 1000 onwards, Western Europe saw the last of the barbarian invasions and became more politically organized. The Vikings had settled in Britain, Ireland, France and elsewhere, whilst Norse Christian kingdoms were developing in their Scandinavian homelands. The Magyars had ceased their expansion in the 10th century, and by the year 1000, the Roman Catholic Apostolic Kingdom of Hungary was recognised in central Europe. With the brief exception of the Mongol invasions, major barbarian incursions ceased. Bulgarian sovereignty was re-established with the anti-Byzantine uprising of the Bulgarians and Vlachs in 1185. The crusaders invaded the Byzantine empire, captured Constantinople in 1204 and established their Latin Empire. Kaloyan of Bulgaria defeated Baldwin I, Latin Emperor of Constantinople, in the Battle of Adrianople on 14 April 1205. The reign of Ivan Asen II of Bulgaria led to maximum territorial expansion and that of Ivan Alexander of Bulgaria to a Second Golden Age of Bulgarian culture. The Byzantine Empire was fully re-established in 1261. In the 11th century, populations north of the Alps began to settle new lands, some of which had reverted to wilderness after the end of the Roman Empire. In what is known as the "great clearances", vast forests and marshes of Europe were cleared and cultivated. At the same time settlements moved beyond the traditional boundaries of the Frankish Empire to new frontiers in Europe, beyond the Elbe river, tripling the size of Germany in the process. Crusaders founded European colonies in the Levant, the majority of the Iberian Peninsula was conquered from the Muslims, and the Normans colonised southern Italy, all part of the major population increase and resettlement pattern. The High Middle Ages produced many different forms of intellectual, spiritual and artistic works. The most famous are the great cathedrals as expressions of Gothic architecture, which evolved from Romanesque architecture. This age saw the rise of modern nation-states in Western Europe and the ascent of the famous Italian city-states, such as Florence and Venice. The influential popes of the Catholic Church called volunteer armies from across Europe to a series of Crusades against the Seljuq Turks, who occupied the Holy Land. The rediscovery of the works of Aristotle led Thomas Aquinas and other thinkers to develop the philosophy of Scholasticism. A divided church The Great Schism between the Western (Catholic) and Eastern (Orthodox) Christian Churches was sparked in 1054 by Pope Leo IX asserting authority over three of the seats in the Pentarchy, in Antioch, Jerusalem and Alexandria. Since the mid-8th century, the Byzantine Empire's borders had been shrinking in the face of Islamic expansion. Antioch had been wrested back into Byzantine control by 1045, but the resurgent power of the Roman successors in the West claimed a right and a duty for the lost seats in Asia and Africa. Pope Leo sparked a further dispute by defending the filioque clause in the Nicene Creed which the West had adopted customarily. The Orthodox today state that the XXVIIIth Canon of the Council of Chalcedon explicitly proclaimed the equality of the Bishops of Rome and Constantinople. The Orthodox also state that the Bishop of Rome has authority only over his own diocese and does not have any authority outside his diocese. There were other less significant catalysts for the Schism however, including variance over liturgy. The Schism of Roman Catholic and Orthodox followed centuries of estrangement between the Latin and Greek worlds. Holy wars After the East–West Schism, Western Christianity was adopted by the newly created kingdoms of Central Europe: Poland, Hungary and Bohemia. The Roman Catholic Church developed as a major power, leading to conflicts between the Pope and emperor. The geographic reach of the Roman Catholic Church expanded enormously due to the conversions of pagan kings (Scandinavia, Lithuania, Poland, Hungary), the Christian Reconquista of Al-Andalus, and the crusades. Most of Europe was Roman Catholic in the 15th century. Early signs of the rebirth of civilization in western Europe began to appear in the 11th century as trade started again in Italy, leading to the economic and cultural growth of independent city-states such as Venice and Florence; at the same time, nation-states began to take form in places such as France, England, Spain, and Portugal, although the process of their formation (usually marked by rivalry between the monarchy, the aristocratic feudal lords and the church) actually took several centuries. These new nation-states began writing in their own cultural vernaculars, instead of the traditional Latin. Notable figures of this movement would include Dante Alighieri and Christine de Pizan (born Christina da Pizzano), the former writing in Italian, and the latter, although an Italian (Venice), relocated to France, writing in French. (See Reconquista for the latter two countries.) Elsewhere, the Holy Roman Empire, essentially based in Germany and Italy, further fragmented into a myriad of feudal principalities or small city states, whose subjection to the emperor was only formal. The 14th century, when the Mongol Empire came to power, is often called the Age of the Mongols. Mongol armies expanded westward under the command of Batu Khan. Their western conquests included almost all of Russia (save Novgorod, which became a vassal), and the Kipchak-Cuman Confederation. Bulgaria, Hungary, and Poland managed to remain sovereign states. Mongolian records indicate that Batu Khan was planning a complete conquest of the remaining European powers, beginning with a winter attack on Austria, Italy and Germany, when he was recalled to Mongolia upon the death of Great Khan Ögedei. Most historians believe only his death prevented the complete conquest of Europe. The areas of Eastern Europe and most of Central Asia that were under direct Mongol rule became known as the Golden Horde. Under Uzbeg Khan, Islam became the official religion of the region in the early 14th century. The invading Mongols, together with their mostly Turkic subjects, were known as Tatars. In Russia, the Tatars ruled the various states of the Rus' through vassalage for over 300 years. In the Northern Europe, Konrad of Masovia gave Chełmno to the Teutonic Knights in 1226 as a base for a Crusade against the Old Prussians and Grand Duchy of Lithuania. The Livonian Brothers of the Sword were defeated by the Lithuanians, so in 1237 Gregory IX merged the remainder of the order into the Teutonic Order as the Livonian Order. By the middle of the century, the Teutonic Knights completed their conquest of the Prussians before converting the Lithuanians in the subsequent decades. The order also came into conflict with the Eastern Orthodox Church of the Pskov and Novgorod Republics. In 1240 the Orthodox Novgorod army defeated the Catholic Swedes in the Battle of the Neva, and, two years later, they defeated the Livonian Order in the Battle on the Ice. The Union of Krewo in 1386, bringing two major changes in the history of the Grand Duchy of Lithuania: conversion to Catholicism and establishment of a dynastic union between the Grand Duchy of Lithuania and the Crown of the Kingdom of Poland marked both the greatest territorial expansion of the Grand Duchy and the defeat of the Teutonic Knights in the Battle of Grunwald in 1410. Late Middle Ages The Late Middle Ages spanned around the 14th and late 15th centuries. Around 1300, centuries of European prosperity and growth came to a halt. A series of famines and plagues, such as the Great Famine of 1315–1317 and the Black Death, killed people in a matter of days, reducing the population of some areas by half as many survivors fled. Kishlansky reports: The Black Death touched every aspect of life, hastening a process of social, economic, and cultural transformation already underway.... Fields were abandoned, workplaces stood idle, international trade was suspended. Traditional bonds of kinship, village, and even religion were broken amid the horrors of death, flight, and failed expectations. "People cared no more for dead men than we care for dead goats," wrote one survivor. Depopulation caused labor to become scarcer; the survivors were better paid and peasants could drop some of the burdens of feudalism. There was also social unrest; France and England experienced serious peasant risings including the Jacquerie and the Peasants' Revolt. At the same time, the unity of the Catholic Church was shattered by the Great Schism. Collectively these events have been called the Crisis of the Late Middle Ages. Beginning in the 14th century, the Baltic Sea became one of the most important trade routes. The Hanseatic League, an alliance of trading cities, facilitated the absorption of vast areas of Poland, Lithuania, and Livonia into trade with other European countries. This fed the growth of powerful states in this part of Europe including Poland–Lithuania, Hungary, Bohemia, and Muscovy later on. The conventional end of the Middle Ages is usually associated with the fall of the city of Constantinople and of the Byzantine Empire to the Ottoman Turks in 1453. The Turks made the city the capital of their Ottoman Empire, which lasted until 1922 and included Egypt, Syria, and most of the Balkans. The Ottoman wars in Europe, also sometimes referred to as the Turkish wars, marked an essential part of the history of the continent as a whole. A key 15th-century development was the advent of the movable type of printing press circa 1439 in Mainz, building upon the impetus provided by the prior introduction of paper from China via the Arabs in the High Middle Ages. Paper was already readily available in Europe by the late 14th century. While forms of moveable type of printing press had been already used in China and Korea, the technique was singularly successful in Europe given the small number of characters of the Latin alphabet, massively reducing costs of book production. The adoption of the technology across the continent at dazzling speed for the remaining part of the 15th century would usher a revolution and by 1500 over 200 cities in Europe had presses that printed between 8 and 20 million books. Homicide rates plunge over 800 years At the local level, levels of violence were extremely high by modern standards in medieval and early modern Europe. They actually reached their peak during the late Middle Ages, having increased since the third century. Typically, small groups would battle their neighbors, using the farm tools at hand such as knives, sickles, hammers and axes. Mayhem and death were deliberate. The vast majority of people lived in rural areas. Cities were few, and small in size, but their concentration of population was conducive to violence. Long-term studies of places such as Amsterdam, Stockholm, Venice and Zurich show the same trends as rural areas. Across Europe, homicide trends (not including military actions) show a steady long-term decline. Regional differences were small, except that Italy's decline was later and slower. From approximately 1200 AD through 1800 AD, homicide rates from violent local episodes declined by a factor of ten, from approximately 32 deaths per 100 000 people to 3.2 per 100 000. In the 20th century the homicide rate fell to 1.4 per 100 000. Police forces seldom existed outside the cities; prisons only became common after 1800. Before then harsh penalties were imposed for homicide (severe whipping or execution) but they proved ineffective at controlling or reducing the insults to honor that precipitated most of the violence. The decline does not correlate with economics. Most historians attribute the trend in homicides to a steady increase in self-control of the sort promoted by Protestantism, and necessitated by schools and factories. Historian Manuel Eisner has summarized the patterns from over 300 historical studies. Early modern Europe The Early Modern period spans the centuries between the Middle Ages and the Industrial Revolution, roughly from 1500 to 1800, or from the discovery of the New World in 1492 to the French Revolution in 1789. The period is characterised by the rise in importance of science and increasingly rapid technological progress, secularised civic politics, and the nation state. Capitalist economies began their rise, and the early modern period also saw the rise and dominance of the economic theory of mercantilism. As such, the early modern period represents the decline and eventual disappearance, in much of the European sphere, of feudalism, serfdom and the power of the Catholic Church. The period includes the Renaissance, the Scientific Revolution, the Protestant Reformation, the disastrous Thirty Years' War, the European colonisation of the Americas and the European witch-hunts. Renaissance Despite these crises, the 14th century was also a time of great progress within the arts and sciences. A renewed interest in ancient Greek and Roman led to the Italian Renaissance. The Renaissance was a cultural movement that profoundly affected European intellectual life in the early modern period. Beginning in Italy, and spreading to the north, west and middle Europe during a cultural lag of some two and a half centuries, its influence affected literature, philosophy, art, politics, science, history, religion, and other aspects of intellectual inquiry. The Italian Petrarch (Francesco Petrarca), deemed the first full-blooded Humanist, wrote in the 1330s: "I am alive now, yet I would rather have been born in another time." He was enthusiastic about Greek and Roman antiquity. In the 15th and 16th centuries the continuing enthusiasm for the ancients was reinforced by the feeling that the inherited culture was dissolving and here was a storehouse of ideas and attitudes with which to rebuild. Matteo Palmieri wrote in the 1430s: "Now indeed may every thoughtful spirit thank god that it has been permitted to him to be born in a new age." The renaissance was born: a new age where learning was very important. The Renaissance was inspired by the growth in the study of Latin and Greek texts and the admiration of the Greco-Roman era as a golden age. This prompted many artists and writers to begin drawing from Roman and Greek examples for their works, but there was also much innovation in this period, especially by multi-faceted artists such as Leonardo da Vinci. The Humanists saw their repossession of a great past as a Renaissance – a rebirth of civilization itself. Important political precedents were also set in this period. Niccolò Machiavelli's political writing in The Prince influenced later absolutism and realpolitik. Also important were the many patrons who ruled states and used the artistry of the Renaissance as a sign of their power. In all, the Renaissance could be viewed as an attempt by intellectuals to study and improve the secular and worldly, both through the revival of ideas from antiquity and through novel approaches to thought – the immediate past being too "Gothic" in language, thought and sensibility. The Scientific Revolution took place in Europe starting towards the second half of the Renaissance period, with the 1543 Nicolaus Copernicus publication De revolutionibus orbium coelestium (On the Revolutions of the Heavenly Spheres) often cited as its beginning. Exploration and trade Toward the end of the period, an era of discovery began. The growth of the Ottoman Empire, culminating in the fall of Constantinople in 1453, cut off trading possibilities with the east. Western Europe was forced to discover new trading routes, as happened with Columbus' travel to the Americas in 1492, and Vasco da Gama's circumnavigation of India and Africa in 1498. The numerous wars did not prevent European states from exploring and conquering wide portions of the world, from Africa to Asia and the newly discovered Americas. In the 15th century, Portugal led the way in geographical exploration along the coast of Africa in search of a maritime route to India, followed by Spain near the close of the 15th century, dividing their exploration of the world according to the Treaty of Tordesillas in 1494. They were the first states to set up colonies in America and European trading posts (factories) along the shores of Africa and Asia, establishing the first direct European diplomatic contacts with Southeast Asian states in 1511, China in 1513 and Japan in 1542. In 1552, Russian tsar Ivan the Terrible conquered two major Tatar khanates, the Khanate of Kazan and the Astrakhan Khanate. The Yermak's voyage of 1580 led to the annexation of the Tatar Siberian Khanate into Russia, and the Russians would soon after conquer the rest of Siberia, steadily expanding to the east and south over the next centuries. Oceanic explorations soon followed by France, England and the Netherlands, who explored the Portuguese and Spanish trade routes into the Pacific Ocean, reaching Australia in 1606 and New Zealand in 1642. Reformation With the development of the printing press, new ideas spread throughout Europe and challenged traditional doctrines in science and theology. Simultaneously, the Protestant Reformation under German Martin Luther questioned Papal authority. The most common dating of the Reformation begins in 1517, when Luther published The Ninety-Five Theses, and concludes in 1648 with the Treaty of Westphalia that ended years of European religious wars. During this period corruption in the Catholic Church led to a sharp backlash in the Protestant Reformation. It gained many followers especially among princes and kings seeking a stronger state by ending the influence of the Catholic Church. Figures other than Martin Luther began to emerge as well like John Calvin whose Calvinism had influence in many countries and King Henry VIII of England who broke away from the Catholic Church in England and set up the Anglican Church; his daughter Queen Elizabeth finished the organization of the church. These religious divisions brought on a wave of wars inspired and driven by religion but also by the ambitious monarchs in Western Europe who were becoming more centralized and powerful. The Protestant Reformation also led to a strong reform movement in the Catholic Church called the Counter-Reformation, which aimed to reduce corruption as well as to improve and strengthen Catholic dogma. Two important groups in the Catholic Church who emerged from this movement were the Jesuits, who helped keep Spain, Portugal, Poland, and other European countries within the Catholic fold, and the Oratorians of Saint Philip Neri, who ministered to the faithful in Rome, restoring their confidence in the Church of Jesus Christ that subsisted substantially in the Church of Rome. Still, the Catholic Church was somewhat weakened by the Reformation, portions of Europe were no longer under its sway and kings in the remaining Catholic countries began to take control of the church institutions within their kingdoms. Unlike many European countries, the Polish–Lithuanian Commonwealth and Hungary were more tolerant. While still enforcing the predominance of Catholicism, they continued to allow the large religious minorities to maintain their faiths, traditions and customs. The Polish–Lithuanian Commonwealth became divided among Catholics, Protestants, Orthodox, Jews and a small Muslim population. Another development was the idea of 'European superiority'. The ideal of civilization was taken over from the ancient Greeks and Romans: Discipline, education and living in the city were required to make people civilized; Europeans and non-Europeans were judged for their civility, and Europe regarded itself as superior to other continents. There was a movement by some such as Montaigne that regarded the non-Europeans as a better, more natural and primitive people. Post services were founded all over Europe, which allowed a humanistic interconnected network of intellectuals across Europe, despite religious divisions. However, the Roman Catholic Church banned many leading scientific works; this led to an intellectual advantage for Protestant countries, where the banning of books was regionally organised. Francis Bacon and other advocates of science tried to create unity in Europe by focusing on the unity in nature.1 In the 15th century, at the end of the Middle Ages, powerful sovereign states were appearing, built by the New Monarchs who were centralising power in France, England, and Spain. On the other hand, the Parliament in the Polish–Lithuanian Commonwealth grew in power, taking legislative rights from the Polish king. The new state power was contested by parliaments in other countries especially England. New kinds of states emerged which were co-operation agreements among territorial rulers, cities, farmer republics and knights. Mercantilism and colonial expansion The Iberian kingdoms were able to dominate colonial activity in the 16th century. The Portuguese forged the first global empire in the 15th and 16th century, whilst during the 16th century and the first half of the 17th century, the crown of Castile (and the overarching Hispanic Monarchy, including Portugal from 1580 to 1640) became the most powerful empire in the world. Spanish dominance in America was increasingly challenged by British, French, Dutch and Swedish colonial efforts of the 17th and 18th centuries. New forms of trade and expanding horizons made new forms of government, law and economics necessary. Colonial expansion continued in the following centuries (with some setbacks, such as successful wars of independence in the British American colonies and then later Haiti, Mexico, Argentina, Brazil, and others amid European turmoil of the Napoleonic Wars). Spain had control of a large part of North America, all of Central America and a great part of South America, the Caribbean and the Philippines; Britain took the whole of Australia and New Zealand, most of India, and large parts of Africa and North America; France held parts of Canada and India (nearly all of which was lost to Britain in 1763), Indochina, large parts of Africa and the Caribbean islands; the Netherlands gained the East Indies (now Indonesia) and islands in the Caribbean; Portugal obtained Brazil and several territories in Africa and Asia; and later, powers such as Germany, Belgium, Italy and Russia acquired further colonies. This expansion helped the economy of the countries owning them. Trade flourished, because of the minor stability of the empires. By the late 16th century, American silver accounted for one-fifth of Spain's total budget. The French colony of Saint-Domingue was one of richest European colonies in the 18th century, operating on a plantation economy fueled by slave labor. During the period of French rule, cash crops produced in Saint-Domingue comprised thirty percent of total French trade while its sugar exports represented forty percent of the Atlantic market. Crisis of the 17th century The 17th century was an era of crisis. Many historians have rejected the idea, while others promote it as an invaluable insight into the warfare, politics, economics, and even art. The Thirty Years' War (1618–1648) focused attention on the massive horrors that wars could bring to entire populations. The 1640s in particular saw more state breakdowns around the world than any previous or subsequent period. The Polish–Lithuanian Commonwealth, the largest state in Europe, temporarily disappeared. In addition, there were secessions and upheavals in several parts of the Spanish empire, the world's first global empire. In Britain the entire Stuart monarchy (England, Scotland, Ireland, and its North American colonies) rebelled. Political insurgency and a spate of popular revolts seldom equalled shook the foundations of most states in Europe and Asia. More wars took place around the world in the mid-17th century than in almost any other period of recorded history. The crises spread far beyond Europe – for example Ming China, the most populous state in the world, collapsed. Across the Northern Hemisphere, the mid-17th century experienced almost unprecedented death rates. Geoffrey Parker, a British historian, suggests that environmental factors may have been in part to blame, especially global cooling. Age of absolutism The "absolute" rule of powerful monarchs such as Louis XIV (ruled France 1643–1715), Peter the Great (ruled Russia 1682–1725), Maria Theresa (ruled Habsburg lands 1740–1780) and Frederick the Great (ruled Prussia 1740–86), produced powerful centralized states, with strong armies and powerful bureaucracies, all under the control of the king. Throughout the early part of this period, capitalism (through mercantilism) was replacing feudalism as the principal form of economic organisation, at least in the western half of Europe. The expanding colonial frontiers resulted in a Commercial Revolution. The period is noted for the rise of modern science and the application of its findings to technological improvements, which animated the Industrial Revolution after 1750. The Reformation had profound effects on the unity of Europe. Not only were nations divided one from another by their religious orientation, but some states were torn apart internally by religious strife, avidly fostered by their external enemies. France suffered this fate in the 16th century in the series of conflicts known as the French Wars of Religion, which ended in the triumph of the Bourbon Dynasty. England avoided this fate for a while and settled down under Elizabeth I to a moderate Anglicanism. Much of modern-day Germany was made up of numerous small sovereign states under the theoretical framework of the Holy Roman Empire, which was further divided along internally drawn sectarian lines. The Polish–Lithuanian Commonwealth is notable in this time for its religious indifference and a general immunity to the horrors of European religious strife. Thirty Years' War 1618–1648 The Thirty Years' War was fought between 1618 and 1648, across Germany and neighbouring areas, and involved most of the major European powers except England and Russia. Beginning as a religious conflict between Protestants and Catholics in Bohemia, it quickly developed into a general war involving Catholics versus Protestants for the most part. The major impact of the war, in which mercenary armies were extensively used, was the devastation of entire regions scavenged bare by the foraging armies. Episodes of widespread famine and disease, and the breakup of family life, devastated the population of the German states and, to a lesser extent, the Low Countries, the Crown of Bohemia and northern parts of Italy, while bankrupting many of the regional powers involved. Between one-fourth and one-third of the German population perished from direct military causes or from disease and starvation, as well as postponed births. After the Peace of Westphalia, which ended the war in favour of nations deciding their own religious allegiance, absolutism became the norm of the continent, while parts of Europe experimented with constitutions foreshadowed by the English Civil War and particularly the Glorious Revolution. European military conflict did not cease, but had less disruptive effects on the lives of Europeans. In the advanced northwest, the Enlightenment gave a philosophical underpinning to the new outlook, and the continued spread of literacy, made possible by the printing press, created new secular forces in thought. From the Union of Krewo (see above) central and eastern Europe was dominated by Kingdom of Poland and Grand Duchy of Lithuania. In the 16th and 17th centuries Central and Eastern Europe was an arena of conflict for domination of the continent between Sweden, the Polish–Lithuanian Commonwealth (involved in series of wars, like Khmelnytsky uprising, Russo-Polish War, the Deluge, etc.) and the Ottoman Empire. This period saw a gradual decline of these three powers which were eventually replaced by new enlightened absolutist monarchies: Russia, Prussia and Austria (the Habsburg monarchy). By the turn of the 19th century they had become new powers, having divided Poland between themselves, with Sweden and Turkey having experienced substantial territorial losses to Russia and Austria respectively as well as pauperisation. War of the Spanish Succession The War of the Spanish Succession (1701–1715) was a major war with France opposed by a coalition of England, the Netherlands, the Habsburg monarchy, and Prussia. Duke of Marlborough commanded the English and Dutch victory at the Battle of Blenheim in 1704. The main issue was whether France under King Louis XIV would take control of Spain's very extensive possessions and thereby become by far the dominant power, or be forced to share power with other major nations. After initial allied successes, the long war produced a military stalemate and ended with the Treaty of Utrecht, which was based on a balance of power in Europe. Historian Russell Weigley argues that the many wars almost never accomplished more than they cost. British historian G. M. Trevelyan argues: That Treaty [of Utrecht], which ushered in the stable and characteristic period of Eighteenth-Century civilization, marked the end of danger to Europe from the old French monarchy, and it marked a change of no less significance to the world at large – the maritime, commercial and financial supremacy of Great Britain. Prussia Frederick the Great, king of Prussia 1740–86, modernized the Prussian army, introduced new tactical and strategic concepts, fought mostly successful wars (Silesian Wars, Seven Years' War) and doubled the size of Prussia. Frederick had a rationale based on Enlightenment thought: he fought total wars for limited objectives. The goal was to convince rival kings that it was better to negotiate and make peace than to fight him. Russia Russia fought numerous wars to achieve rapid expansion toward the east – i.e. Siberia, Far East, south, to the Black Sea, and south-east and to central Asia. Russia boasted a large and powerful army, a very large and complex internal bureaucracy, and a splendid court that rivaled Paris and London. However the government was living far beyond its means and seized Church lands, leaving organized religion in a weak condition. Throughout the 18th century Russia remained "a poor, backward, overwhelmingly agricultural, and illiterate country." Enlightenment The Enlightenment was a powerful, widespread cultural movement of intellectuals beginning in late 17th-century Europe emphasizing the power of reason rather than tradition; it was especially favourable to science (especially Isaac Newton's physics) and hostile to religious orthodoxy (especially of the Catholic Church). It sought to analyze and reform society using reason, to challenge ideas grounded in tradition and faith, and to advance knowledge through the scientific method. It promoted scientific thought, skepticism, and intellectual interchange. The Enlightenment was a revolution in human thought. This new way of thinking was that rational thought begins with clearly stated principles, uses correct logic to arrive at conclusions, tests the conclusions against evidence, and then revises the principles in light of the evidence. Enlightenment thinkers opposed superstition. Some Enlightenment thinkers collaborated with Enlightened despots, absolutist rulers who attempted to forcibly impose some of the new ideas about government into practice. The ideas of the Enlightenment exerted significant influence on the culture, politics, and governments of Europe. Originating in the 17th century, it was sparked by philosophers Francis Bacon (1562–1626), Baruch Spinoza (1632–1677), John Locke (1632–1704), Pierre Bayle (1647–1706), Voltaire (1694–1778), Francis Hutcheson (1694–1746), David Hume (1711–1776) and physicist Isaac Newton (1643–1727). Ruling princes often endorsed and fostered these figures and even attempted to apply their ideas of government in what was known as enlightened absolutism. The Scientific Revolution is closely tied to the Enlightenment, as its discoveries overturned many traditional concepts and introduced new perspectives on nature and man's place within it. The Enlightenment flourished until about 1790–1800, at which point the Enlightenment, with its emphasis on reason, gave way to Romanticism, which placed a new emphasis on emotion; a Counter-Enlightenment began to increase in prominence. The Romantics argued that the Enlightenment was reductionistic insofar as it had largely ignored the forces of imagination, mystery, and sentiment. In France, Enlightenment was based in the salons and culminated in the great Encyclopédie (1751–72) edited by Denis Diderot (1713–1784) and (until 1759) Jean le Rond d'Alembert (1717–1783) with contributions by hundreds of leading intellectuals who were called philosophes, notably Voltaire (1694–1778), Rousseau (1712–1778) and Montesquieu (1689–1755). Some 25,000 copies of the 35 volume encyclopedia were sold, half of them outside France. These new intellectual strains would spread to urban centres across Europe, notably England, Scotland, the German states, the Netherlands, Poland, Russia, Italy, Austria, and Spain, as well as Britain's American colonies. The political ideals of the Enlightenment influenced the American Declaration of Independence, the United States Bill of Rights, the French Declaration of the Rights of Man and of the Citizen, and the Polish–Lithuanian Constitution of 3 May 1791. Taking a long-term historical perspective, Norman Davies has argued that Freemasonry was a powerful force on behalf of Liberalism and Enlightenment ideas in Europe, from about 1700 to the 20th century. It expanded rapidly during the Age of Enlightenment, reaching practically every country in Europe. Prominent members included Montesquieu, Voltaire, Sir Robert Walpole, Wolfgang Amadeus Mozart, Johann Wolfgang von Goethe, Benjamin Franklin, and George Washington. Steven C. Bullock notes that in the late 18th century, English lodges were headed by the Prince of Wales, Prussian lodges by king Frederick the Great, and French lodges by royal princes. Emperor Napoleon selected as Grand Master of France his own brother. The great enemy of Freemasonry was the Roman Catholic Church, so that in countries with a large Catholic element, such as France, Italy, Austria, Spain and Mexico, much of the ferocity of the political battles involve the confrontation between supporters of the Church versus active Masons. 20th-century totalitarian and revolutionary movements, especially the Fascists and Communists, crushed the Freemasons. From revolution to imperialism (1789–1914) The "long 19th century", from 1789 to 1914 saw the drastic social, political and economic changes initiated by the Industrial Revolution, the French Revolution and the Napoleonic Wars. Following the reorganisation of the political map of Europe at the Congress of Vienna in 1815, Europe experienced the rise of Nationalism, the rise of the Russian Empire and the peak of the British Empire, as well as the decline of the Ottoman Empire. Finally, the rise of the German Empire and the Austro-Hungarian Empire initiated the course of events that culminated in the outbreak of the First World War in 1914. Industrial Revolution The Industrial Revolution was a period in the late 18th century and early 19th century when major changes in agriculture, manufacturing, and transport impacted Britain and subsequently spread to the United States and Western Europe, a process that continues as industrialisation. Technological advancements, most notably the utilization of the steam engine, were major catalysts in the industrialisation process. It started in England and Scotland in the mid-18th century with the mechanisation of the textile industries, the development of iron-making techniques and the increased use of coal as the main fuel. Trade expansion was enabled by the introduction of canals, improved roads and railways. The introduction of steam power (fuelled primarily by coal) and powered machinery (mainly in textile manufacturing) underpinned the dramatic increases in production capacity. The development of all-metal machine tools in the first two decades of the 19th century facilitated the manufacture of more production machines for manufacturing in other industries. The effects spread throughout Western Europe and North America during the 19th century, eventually affecting most of the world. The impact of this change on society was enormous. Era of the French Revolution Historians R.R. Palmer and Joel Colton argue: In 1789 France fell into revolution, and the world has never since been the same. The French Revolution was by far the most momentous upheaval of the whole revolutionary age. It replaced the "old regime" with "modern society," and at its extreme phase became very radical, so much so that all later revolutionary movements have looked back to it as a predecessor to themselves.... From the 1760s to 1848, the role of France was decisive. The era of the French Revolution and the subsequent Napoleonic wars was a difficult time for monarchs. Tsar Paul I of Russia was assassinated; King Louis XVI of France was executed, as was his queen Marie Antoinette. Furthermore, kings Charles IV of Spain, Ferdinand VII of Spain and Gustav IV Adolf of Sweden were deposed as were ultimately the Emperor Napoleon and all of the relatives he had installed on various European thrones. King Frederick William III of Prussia and Emperor Francis II of Austria barely clung to their thrones. King George III of Great Britain lost the better part of the First British Empire. The American Revolution (1775–1783) was the first successful revolt of a colony against a European power. It proclaimed, in the words of Thomas Jefferson, that "all men are created equal," a position based on the principles of the Enlightenment. It rejected aristocracy and established a republican form of government under George Washington that attracted worldwide attention. The French Revolution (1789–1804) was a product of the same democratic forces in the Atlantic World and had an even greater impact. French historian François Aulard says: From the social point of view, the Revolution consisted in the suppression of what was called the feudal system, in the emancipation of the individual, in greater division of landed property, the abolition of the privileges of noble birth, the establishment of equality, the simplification of life.... The French Revolution differed from other revolutions in being not merely national, for it aimed at benefiting all humanity." French intervention in the American Revolutionary War had nearly bankrupted the state. After repeated failed attempts at financial reform, King Louis XVI had to convene the Estates-General, a representative body of the country made up of three estates: the clergy, the nobility, and the commoners. The third estate, joined by members of the other two, declared itself to be a National Assembly and swore an oath not to dissolve until France had a constitution and created, in July, the National Constituent Assembly. At the same time the people of Paris revolted, famously storming the Bastille prison on 14 July 1789. At the time the assembly wanted to create a constitutional monarchy, and over the following two years passed various laws including the Declaration of the Rights of Man and of the Citizen, the abolition of feudalism, and a fundamental change in the relationship between France and Rome. At first the king agreed with these changes and enjoyed reasonable popularity with the people. As anti-royalism increased along with threat of foreign invasion, the king tried to flee and join France's enemies. He was captured and on 21 January 1793, having been convicted of treason, he was guillotined. On 20 September 1792 the National Convention abolished the monarchy and declared France a republic. Due to the emergency of war, the National Convention created the Committee of Public Safety, controlled by Maximilien de Robespierre of the Jacobin Club, to act as the country's executive. Under Robespierre, the committee initiated the Reign of Terror, during which up to 40,000 people were executed in Paris, mainly nobles and those convicted by the Revolutionary Tribunal, often on the flimsiest of evidence. Internal tensions at Paris drove the Committee towards increasing assertions of radicalism and increasing suspicions, fueling new terror: A few months into this phase, more and more prominent revolutionaries were being sent to the guillotine by Robespierre and his faction, for example Madame Roland and Georges Danton. Elsewhere in the country, counter-revolutionary insurrections were brutally suppressed. The regime was overthrown in the coup of 9 Thermidor (27 July 1794) and Robespierre was executed. The regime which followed ended the Terror and relaxed Robespierre's more extreme policies. Napoleon Napoleon Bonaparte was one of the world's most famous soldiers and statesmen, leading France to great victories over numerous European enemies. Despite modest origins he became Emperor and restructured much of European diplomacy, politics and law, until he was forced to abdicate in 1814. His 100-day comeback in 1815 failed at the Battle of Waterloo, and he died in exile on a remote island, remembered as a great hero by many Frenchmen and as a great villain by British and other enemies. Napoleon, despite his youth, was France's most successful general in the Revolutionary wars, having conquered large parts of Italy and forced the Austrians to sue for peace. In 1799 on 18 Brumaire (9 November) he overthrew the feeble government, replacing it with the Consulate, which he dominated. He gained popularity in France by restoring the Church, keeping taxes low, centralizing power in Paris, and winning glory on the battlefield. In 1804 he crowned himself Emperor. In 1805, Napoleon planned to invade Britain, but a renewed British alliance with Russia and Austria (Third Coalition), forced him to turn his attention towards the continent, while at the same time the French fleet was demolished by the British at the Battle of Trafalgar, ending any plan to invade Britain. On 2 December 1805, Napoleon defeated a numerically superior Austro-Russian army at Austerlitz, forcing Austria's withdrawal from the coalition (see Treaty of Pressburg) and dissolving the Holy Roman Empire. In 1806, a Fourth Coalition was set up. On 14 October Napoleon defeated the Prussians at the Battle of Jena-Auerstedt, marched through Germany and defeated the Russians on 14 June 1807 at Friedland. The Treaties of Tilsit divided Europe between France and Russia and created the Duchy of Warsaw. On 12 June 1812 Napoleon invaded Russia with a Grande Armée of nearly 700,000 troops. After the measured victories at Smolensk and Borodino Napoleon occupied Moscow, only to find it burned by the retreating Russian army. He was forced to withdraw. On the march back his army was harassed by Cossacks, and suffered disease and starvation. Only 20,000 of his men survived the campaign. By 1813 the tide had begun to turn from Napoleon. Having been defeated by a seven nation army at the Battle of Leipzig in October 1813, he was forced to abdicate after the Six Days' Campaign and the occupation of Paris. Under the Treaty of Fontainebleau he was exiled to the island of Elba. He returned to France on 1 March 1815 (see Hundred Days), raised an army, but was finally defeated by a British and Prussian force at the Battle of Waterloo on 18 June 1815 and exiled to the small British island of Saint Helena in the South Atlantic. Impact of the French Revolution Roberts finds that the Revolutionary and Napoleonic wars, from 1793 to 1815, caused 4 million deaths (of whom 1 million were civilians); 1.4 million were French deaths. Outside France the Revolution had a major impact. Its ideas became widespread. Roberts argues that Napoleon was responsible for key ideas of the modern world, so that, "meritocracy, equality before the law, property rights, religious toleration, modern secular education, sound finances, and so on-were protected, consolidated, codified, and geographically extended by Napoleon during his 16 years of power." Furthermore, the French armies in the 1790s and 1800s directly overthrew feudal remains in much of western Europe. They liberalised property laws, ended seigneurial dues, abolished the guild of merchants and craftsmen to facilitate entrepreneurship, legalised divorce, closed the Jewish ghettos and made Jews equal to everyone else. The Inquisition ended as did the Holy Roman Empire. The power of church courts and religious authority was sharply reduced and equality under the law was proclaimed for all men. In foreign affairs, the French Army down to 1812 was quite successful. Roberts says that Napoleon fought 60 battles, losing only seven. France conquered Belgium and turned it into another province of France. It conquered the Netherlands, and made it a client state. It took control of the German areas on the left bank of the Rhine River and set up a puppet Confederation of the Rhine. It conquered Switzerland and most of Italy, setting up a series of puppet states. The result was glory for France, and an infusion of much needed money from the conquered lands, which also provided direct support to the French Army. However the enemies of France, led by Britain and funded by the inexhaustible British Treasury, formed a Second Coalition in 1799 (with Britain joined by Russia, the Ottoman Empire and Austria). It scored a series of victories that rolled back French successes, and trapped the French Army in Egypt. Napoleon himself slipped through the British blockade in October 1799, returning to Paris, where he overthrew the government and made himself the ruler. Napoleon conquered most of Italy in the name of the French Revolution in 1797–99. He consolidated old units and split up Austria's holdings. He set up a series of new republics, complete with new codes of law and abolition of old feudal privileges. Napoleon's Cisalpine Republic was centered on Milan; Genoa became a republic; the Roman Republic was formed as well as the small Ligurian Republic around Genoa. The Neapolitan Republic was formed around Naples, but it lasted only five months. He later formed the Kingdom of Italy, with his brother as King. In addition, France turned the Netherlands into the Batavian Republic, and Switzerland into the Helvetic Republic. All these new countries were satellites of France, and had to pay large subsidies to Paris, as well as provide military support for Napoleon's wars. Their political and administrative systems were modernized, the metric system introduced, and trade barriers reduced. Jewish ghettos were abolished. Belgium and Piedmont became integral parts of France. Most of the new nations were abolished and returned to prewar owners in 1814. However, Artz emphasizes the benefits the Italians gained from the French Revolution: For nearly two decades the Italians had excellent codes of law, a fair system of taxation, a better economic situation, and more religious and intellectual toleration than they had known for centuries.... Everywhere old physical, economic, and intellectual barriers had been thrown down and the Italians had begun to be aware of a common nationality. Likewise in Switzerland the long-term impact of the French Revolution has been assessed by Martin: It proclaimed the equality of citizens before the law, equality of languages, freedom of thought and faith; it created a Swiss citizenship, basis of our modern nationality, and the separation of powers, of which the old regime had no conception; it suppressed internal tariffs and other economic restraints; it unified weights and measures, reformed civil and penal law, authorized mixed marriages (between Catholics and Protestants), suppressed torture and improved justice; it developed education and public works. The greatest impact came of course in France itself. In addition to effects similar to those in Italy and Switzerland, France saw the introduction of the principle of legal equality, and the downgrading of the once powerful and rich Catholic Church to just a bureau controlled by the government. Power became centralized in Paris, with its strong bureaucracy and an army supplied by conscripting all young men. French politics were permanently polarized – new names were given, "left" and "right" for the supporters and opponents of the principles of the Revolution. British historian Max Hastings says there is no question that as a military genius Napoleon ranks with Alexander the Great and Julius Caesar in greatness. However, in the political realm, historians debate whether Napoleon was "an enlightened despot who laid the foundations of modern Europe or, instead, a megalomaniac who wrought greater misery than any man before the coming of Hitler". Religion By the 19th century, governments increasingly took over traditional religious roles, paying much more attention to efficiency and uniformity than to religiosity. Secular bodies took control of education away from the churches, abolished taxes and tithes for the support of established religions, and excluded bishops from the upper houses. Secular laws increasingly regulated marriage and divorce, and maintaining birth and death registers became the duty of local officials. Although the numerous religious denominations in the United States founded many colleges and universities, that was almost exclusively a state function across Europe. Imperial powers protected Christian missionaries in African and Asian colonies. In France and other largely Catholic nations, anti-clerical political movements tried to reduce the role of the Catholic Church. Likewise briefly in Germany in the 1870s there was a fierce Kulturkampf (culture war) against Catholics, but the Catholics successfully fought back. The Catholic Church concentrated more power in the papacy and fought against secularism and socialism. It sponsored devotional reforms that gained wide support among the churchgoers. Protestantism Historian Kenneth Scott Latourette argues that the outlook for Protestantism at the start of the 19th century was discouraging. It was a regional religion based in Northwestern Europe, with an outpost in the sparsely settled United States. It was closely allied with government, as in Scandinavia, the Netherlands, Prussia, and especially Great Britain. The alliance came at the expense of independence, as the government made the basic policy decisions, down to such details as the salaries of ministers and location of new churches. The dominant intellectual currents of the Enlightenment promoted rationalism, and most Protestant leaders preached a sort of deism. Intellectually, the new methods of historical and anthropological study undermine automatic acceptance of biblical stories, as did the sciences of geology and biology. Industrialization was a strongly negative factor, as workers who moved to the city seldom joined churches. The gap between the church and the unchurched grew rapidly, and secular forces, based both in socialism and liberalism undermine the prestige of religion. Despite the negative forces, Protestantism demonstrated a striking vitality by 1900. Shrugging off Enlightenment rationalism, Protestants embraced romanticism, with the stress on the personal and the invisible. Entirely fresh ideas as expressed by Friedrich Schleiermacher, Soren Kierkegaard, Albrecht Ritschl and Adolf von Harnack restored the intellectual power of theology. There was more attention to historic creeds such as the Augsburg, the Heidelberg, and the Westminster confessions. In England, Anglicans emphasize the historically Catholic components of their heritage, as the High Church element reintroduced vestments and incense into their rituals. The stirrings of pietism on the Continent, and evangelicalism in Britain expanded enormously, leading the devout away from an emphasis on formality and ritual and toward an inner sensibility toward personal relationship to Christ. Social activities, in education and in opposition to social vices such as slavery, alcoholism and poverty provided new opportunities for social service. Above all, worldwide missionary activity became a highly prized goal, proving quite successful in close cooperation with European colonialists, particularly during the New Imperialism period. Nations rising Emerging nationalism The political development of nationalism and the push for popular sovereignty culminated with the ethnic/national revolutions of Europe. During the 19th century nationalism became one of the most significant political and social forces in history; it is typically listed among the top causes of World War I. Napoleon's conquests of the German and Italian states around 1800–1806 played a major role in stimulating nationalism and the demands for national unity. Germany In the German states east of Prussia Napoleon abolished many of the old or medieval relics, such as dissolving the Holy Roman Empire in 1806. He imposed rational legal systems and demonstrated how dramatic changes were possible. For example, his organization of the Confederation of the Rhine in 1806 promoted a feeling of German nationalism. Nationalists sought to encompass masculinity in their quest for strength and unity. In the 1860s it was Prussian chancellor Otto von Bismarck who achieved German unification in 1870 after the many smaller states followed Prussia's leadership in wars against Denmark, Austria and France. Italy Italian nationalism emerged in the 19th century and was the driving force for Italian unification or the "Risorgimento" (meaning the Resurgence or revival). It was the political and intellectual movement that consolidated different states of the Italian Peninsula into the single state of the Kingdom of Italy in 1860. The memory of the Risorgimento is central to both Italian nationalism and Italian historiography. Serbia For centuries the Orthodox Christian Serbs were ruled by the Muslim-controlled Ottoman Empire. The success of the Serbian revolution (1804–1817) against Ottoman rule in 1817 marked the foundation of modern Principality of Serbia. It achieved de facto independence in 1867 and finally gained recognition by the Great Powers in the Berlin Congress of 1878. The Serbs developed a larger vision for nationalism in Pan-Slavism and with Russian support sought to pull the other Slavs out of the Austro-Hungarian Empire. Austria, with German backing, tried to crush Serbia in 1914 but Russia intervened, thus igniting the First World War in which Austria dissolved into nation states. In 1918, the region of Vojvodina proclaimed its secession from Austria-Hungary to unite with the pan-Slavic State of Slovenes, Croats and Serbs; the Kingdom of Serbia joined the union on 1 December 1918, and the country was named Kingdom of Serbs, Croats, and Slovenes. It was renamed Yugoslavia, which was never able to tame the multiple nationalities and religions and it flew apart in civil war in the 1990s. Greece The Greek drive for independence from the Ottoman Empire inspired supporters across Christian Europe, especially in Britain. France, Russia and Britain intervened to make this nationalist dream become reality with the Greek War of Independence (1821-1829/1830). Bulgaria Bulgarian modern nationalism emerged under Ottoman rule in the late 18th and early 19th century, under the influence of western ideas such as liberalism and nationalism, which trickled into the country after the French revolution, mostly via Greece, although there were stirrings in the 18th century. Russia, as a World Great Power of fellow Orthodox Slavs, could appeal to the Bulgarians in a way that Austria could not. An autonomous Bulgarian Exarchate was established in 1870/1872 for the diocese of Bulgaria as well as for those, wherein at least two-thirds of Orthodox Christians were willing to join it. The April Uprising in 1876 indirectly resulted in the re-establishment of Bulgaria in 1878. Poland The cause of Polish nationalism was repeatedly frustrated before 1918. In the 1790s, Germany, Russia and Austria partitioned Poland. Napoleon set up the Duchy of Warsaw, a new Polish state that ignited a spirit of nationalism. Russia took it over in 1815 as Congress Poland with the tsar as King of Poland. Large-scale nationalist revolts erupted in 1830 and 1863–64 but were harshly crushed by Russia, which tried to Russify the Polish language, culture and religion. The collapse of the Russian Empire in the First World War enabled the major powers to reestablish an independent Second Polish Republic, which survived until 1939. Meanwhile, Poles in areas controlled by Germany moved into heavy industry but their religion came under attack by Bismarck in the Kulturkampf of the 1870s. The Poles joined German Catholics in a well-organized new Centre Party, and defeated Bismarck politically. He responded by stopping the harassment and cooperating with the Centre Party. Spain After the War of the Spanish Succession, rooted in the political position of the Count-Duke of Olivares and the absolutism of Philip V, the assimilation of the Crown of Aragon by the Castilian Crown through the Decrees of Nova planta was the first step in the creation of the Spanish nation state. As in other contemporary European states, political union was the first step in the creation of the Spanish nation-state, in this case not on a uniform ethnic basis, but through the imposition of the political and cultural characteristics of the dominant ethnic group, in this case the Castilians, over those of other ethnic groups, who became national minorities to be assimilated. In fact, since the political unification of 1714, Spanish assimilation policies towards Catalan-speaking territories (Catalonia, Valencia, the Balearic Islands, part of Aragon) and other national minorities have been a historical constant. The nationalization process accelerated in the 19th century, in parallel to the origin of Spanish nationalism, the social, political and ideological movement that tried to shape a Spanish national identity based on the Castilian model, in conflict with the other historical nations of the State. These nationalist policies, sometimes very aggressive, and still in force, have been, and still are, the seed of repeated territorial conflicts within the State. Education An important component of nationalism was the study of the nation's heritage, emphasizing the national language and literary culture. This stimulated, and was in turn strongly supported by, the emergence of national educational systems reaching the general population. Latin gave way to the national language, and compulsory education, with strong support from modernizers and the media, became standard in Germany and eventually the other West European nations. Voting reforms extended the franchise to the previously excluded elements. A strong sentiment among the elites was the necessity for compulsory public education, so that the new electorate could understand and handle its duties. Every country developed a sense of national origins – the historical accuracy was less important than the motivation toward patriotism. Universal compulsory education was extended as well to girls, at least at the elementary level. By the 1890s, strong movements emerged in some countries, including France, Germany and the United States, to extend compulsory education to the secondary level. Ideological coalitions After the defeat of revolutionary France, the great powers tried to restore the situation which existed before 1789. In 1815 at the Congress of Vienna, the major powers of Europe managed to produce a peaceful balance of power among the various European empires. This was known as the Metternich system. The powerbase of their support was the aristocracy, with its great landed wealth and control of the government, the church, and the military in most countries. However, their reactionary efforts were unable to stop the spread of revolutionary movements: the middle classes had been deeply influenced by the ideals of the French revolution, and the Industrial Revolution brought important economical and social changes. Radical intellectuals looked to the working classes for a base for socialist, communist and anarchistic ideas. Widely influential was the 1848 pamphlet by Karl Marx and Friedrich Engels The Communist Manifesto. The middle classes and businessmen promoted liberalism, free trade and capitalism. Aristocratic elements concentrated in government service, the military and the established churches. Nationalist movements (in Germany, Italy, Poland, Hungary, and elsewhere) called upon the "racial" unity (which usually meant a common language and an imagined common ethnicity) to seek national unification and/or liberation from foreign rule. As a result, the period between 1815 and 1871 saw a large number of revolutionary attempts and independence wars. Greece successfully revolted against Ottoman rule in the 1820s. European diplomats and intellectuals saw the Greek struggle for independence, with its accounts of Turkish atrocities, in a romantic light. France under Napoleon III Napoleon III, nephew of Napoleon I, parlayed his famous name and to widespread popularity across France. He returned from exile in 1848, promising to stabilize the chaotic political situation. He was elected president and maneuvered successfully to name himself Emperor, a move approved later by a large majority of the French electorate. The first part of his Imperial term brought many important reforms, facilitated by Napoleon's control of the lawmaking body, the government, and the French Armed Forces. Hundreds of old Republican leaders were arrested and deported. Napoleon controlled the media and censored the news. In compensation for the loss of freedom, Napoleon gave the people new hospitals and asylums, beautified and modernized Paris, and built a modern railroad and transportation system that dramatically improved commerce, and helped the many small farmers as well. The economy grew, but industrialization was not as rapid as Britain, and France depended largely on small family-oriented firms as opposed to the large companies that were emerging in the United States and Germany. France was on the winning side in the Crimean War (1854–56), but after 1858 Napoleon's foreign-policy was less and less successful. He antagonized Great Britain and failed to appreciate the danger of war with Prussia. Foreign-policy blunders finally destroyed his reign in 1870–71. He gained worldwide attention for his aggressive foreign policy in Europe, Mexico, and worldwide. He helped in the unification of Italy by fighting the Austrian Empire and joined the Crimean War on the side of the United Kingdom to defend the Ottoman Empire against Russia. His empire collapsed after being defeated in the Franco-Prussian War. France became a republic, but until the 1880s there was a strong popular demand for return to monarchy. That never happened because of the blunders made by the available monarchs. Hostility to the Catholic Church became a major issue, as France battle between secular and religious forces well into the 20th century, with the secular elements usually more successful. The French Third Republic emerged in 1871, was on the winning side of the first world war, and was finally overthrown when it was defeated in 1940 in World War II. Major powers Most European states had become constitutional (rather than absolute) monarchies by 1871, and Germany and Italy merged many small city-states to become united nation-states. Germany in particular increasingly dominated the continent in terms of economics and political power. Meanwhile, on a global scale, Great Britain, with its far-flung British Empire, unmatched Royal Navy, and powerful bankers, became the world's first global power. The sun never set on its territories, while an informal empire operated through British financiers, entrepreneurs, traders and engineers who established operations in many countries, and largely dominated Latin America. The British were especially famous for financing and constructing railways around the world. Bismarck's Germany From his base in Prussia, Otto von Bismarck in the 1860s engineered a series of short, decisive wars, that unified most of the German states (excluding Austria) into a powerful German Empire under Prussian leadership. He humiliated France in the process, but kept on good terms with Austria-Hungary. With that accomplished by 1871 he then skillfully used balance of power diplomacy to preserve Germany's new role and keep Europe at peace. The new German Empire industrialized rapidly and challenged Britain for economic leadership. Bismarck disliked colonies but public and elite opinion forced him to build an overseas empire. He was removed from office in 1890 by an aggressive young Kaiser Wilhelm II, who pursued a disruptive foreign policy that polarized Europe into rival camps. These rival camps went to war with each other in 1914. Austrian and Russian empires The power of nationalism to create new states was irresistible in the 19th century, and the process could lead to collapse in the absence of a strong nationalism. Austria-Hungary had the advantage of size, but multiple disadvantages. There were rivals on four sides, its finances were unstable, the population was fragmented into multiple ethnicities and languages that served as the bases for separatist nationalisms. It had a large army with good forts, but its industrial base was thin. Its naval resources were so minimal that it did not attempt to build an overseas empire. It did have the advantage of good diplomats, typified by Metternich (Foreign Minister 1809–1848, Prime Minister, 1821–1848)). They employed a grand strategy for survival that balanced out different forces, set up buffer zones, and kept the Hapsburg empire going despite wars with the Ottomans, Frederick the Great, Napoleon and Bismarck, until the final disaster of the First World War. The Empire overnight disintegrated into multiple states based on ethnic nationalism and the principle of self-determination. The Russian Empire likewise brought together a multitude of languages and cultures, so that its military defeat in the First World War led to multiple splits that created independent Finland, Latvia, Lithuania, Estonia, and Poland, and for a brief spell, independent Ukraine, Armenia, Georgia, and Azerbaijan. Growth of European armies 1871 to 1904 Table: European armies on active duty in 1871, 1904 Imperialism Colonial empires were the product of the European Age of Discovery from the 15th century. The initial impulse behind these dispersed maritime empires and those that followed was trade, driven by the new ideas and the capitalism that grew out of the Renaissance. Both the Portuguese Empire and Spanish Empire quickly grew into the first global political and economic systems with territories spread around the world. Subsequent major European colonial empires included the French, Dutch, and British empires. The latter, consolidated during the period of British maritime hegemony in the 19th century, became the largest empire in history because of the improved ocean transportation technologies of the time as well as electronic communication through the telegraph, cable, and radio. At its height in 1920, the British Empire covered a quarter of the Earth's land area and comprised a quarter of its population. Other European countries, such as Belgium, Germany, and Italy, pursued colonial empires as well (mostly in Africa), but they were smaller. Ignoring the oceans, Russia built its Russian Empire through conquest by land in Eastern Europe, and Asia. By the mid-19th century, the Ottoman Empire had declined enough to become a target for other global powers (see History of the Balkans). This instigated the Crimean War in 1854 and began a tenser period of minor clashes among the globe-spanning empires of Europe that eventually set the stage for the First World War. In the second half of the 19th century, the Kingdom of Sardinia and the Kingdom of Prussia carried out a series of wars that resulted in the creation of Italy and Germany as nation-states, significantly changing the balance of power in Europe. From 1870, Otto von Bismarck engineered a German hegemony of Europe that put France in a critical situation. It slowly rebuilt its relationships, seeking alliances with Russia and Britain to control the growing power of Germany. In this way, two opposing sides – the Triple Alliance of 1882 (Germany, Austria-Hungary and Italy) and the Triple Entente of 1907 (Britain, France and Russia) – formed in Europe, improving their military forces and alliances year-by-year. 1914–1945: two world wars German-American historian Konrad Jarausch, asked if he agreed that "the European record of the past century [was] just one gigantic catastrophe", argues: It is true that the first half of the 20th century was full of internecine warfare, economic depression, ethnic cleansing and racist genocide that killed tens of millions of people, more than any other period in human history. But looking only at the disasters creates an incomplete perception, because the second half of the century witnessed a much more positive development in spite of the Cold War. After the defeat of Fascism in 1945, the peaceful revolution of 1989/90 also liberated the East from Communist control in a quite unexpected fashion. As a result, Europeans generally live more free, prosperous and healthy lives than ever before. The "short twentieth century", from 1914 to 1991, included the First World War, the Second World War and the Cold War. The First World War used modern technology to kill millions of soldiers. Victory by Britain, France, the United States and other allies drastically changed the map of Europe, ending four major land empires (the Russian, German, Austro-Hungarian and Ottoman empires) and leading to the creation of nation-states across Central and Eastern Europe. The October Revolution in Russia led to the creation of the Soviet Union (1917–1991) and the rise of the international communist movement. Widespread economic prosperity was typical of the period before 1914, and 1920–1929. After the onset of the Great Depression in 1929, however, democracy collapsed in most of Europe. Benito Mussolini's National Fascist Party took control in Italy, and the even more aggressive Nazi Party led by Adolf Hitler took control of Germany, 1933–45. The Second World War was fought on an even larger scale than the First war, killing many more people, and using even more advanced technology. It ended with the division of Europe between East and West, with the East under the control of the Soviet Union and the West dominated by NATO. The two sides engaged in the Cold War, with actual conflict taking place not in Europe but in Asia in the Korean War and the Vietnam War. The Imperial system collapsed. The remaining colonial empires ended through the decolonisation of European rule in Africa and Asia. The fall of Soviet Communism (1989–1991) left the West dominant and enabled the reunification of Germany. It accelerated the process of a European integration to include Eastern Europe. The European Union continues today, but with German economic dominance. Since the worldwide Great Recession of 2008, European growth has been slow, and financial crises have hit Greece and other countries. Modern-day Russia is weaker by military might, compared to when it was a superpower as a part of the Soviet Union, but has retained its historical status as both a great power and a regional power, confronting Ukraine and other post-Soviet states. World War I After the relative peace of most of the 19th century, the rivalry between European powers, compounded by a rising nationalism among ethnic groups, exploded in August 1914, when the First World War started. Over 65 million European soldiers were mobilised from 1914 to 1918; 20 million soldiers and civilians died, and 21 million were seriously wounded. On one side were Germany, Austria-Hungary, the Ottoman Empire and Bulgaria (the Central Powers/Triple Alliance), while on the other side stood Serbia and the Triple Entente – the coalition of France, Britain and Russia, which were joined by Italy in 1915, Romania in 1916 and by the United States in 1917. The Western Front involved especially brutal combat without any territorial gains by either side. Single battles like Verdun and the Somme killed hundreds of thousands of men while leaving the stalemate unchanged. Heavy artillery and machine guns caused most of the casualties, supplemented by poison gas. Czarist Russia collapsed in the February Revolution of 1917 and Germany claimed victory on the Eastern Front. After eight months of liberal rule, the October Revolution brought Vladimir Lenin and the Bolsheviks to power, leading to the creation of the Soviet Union in place of the disintegrated Russian Empire. With American entry into the war in 1917 on the Allied side, and the failure of Germany's spring 1918 offensive, Germany had run out of manpower, while an average of 10,000 American troops were arriving in France every day in the summer of 1918. Germany's allies, Austria-Hungary and the Ottoman Empire, surrendered and dissolved, followed by Germany on 11 November 1918. The victors forced Germany to assume responsibility for the conflict and pay war reparations. One factor in determining the outcome of the war was that the Allies had significantly more economic resources they could spend on the war. One estimate (using 1913 US dollars) is that the Allies spent $58 billion on the war and the Central Powers only $25 billion. Among the Allies, Britain spent $21 billion and the U.S. $17 billion; among the Central Powers Germany spent $20 billion. Paris Peace Conference The world war was settled by the victors at the Paris Peace Conference in 1919. Two dozen nations sent delegations, and there were many nongovernmental groups, but the defeated powers were not invited. The "Big Four" were President Woodrow Wilson of the United States, Prime Minister David Lloyd George of Great Britain, Georges Clemenceau of France, and, of least importance, Italian Prime Minister Vittorio Orlando. Each has a large staff of experts. They met together informally 145 times and made all the major decisions, which in turn were ratified by the others. The major decisions were the creation of the League of Nations; the six peace treaties with defeated enemies, most notable the Treaty of Versailles with Germany; the awarding of German and Ottoman overseas possessions as "mandates", chiefly to Britain and France; and the drawing of new national boundaries (sometimes with plebiscites) to better reflect the forces of nationalism. The Big Four implemented sweeping changes to the political geography of the world. Most famously, the Treaty of Versailles itself weakened Germany's military power and placed full blame for the war and costly reparations on its shoulders – the humiliation and resentment in Germany was probably one of the causes of Nazi success and indirectly a cause of World War II. At the insistence of President Wilson, the Big Four required the Second Polish Republic to sign a treaty on 28 June 1919 that guaranteed minority rights in the new nation. Poland signed under protest, and made little effort to enforce the specified rights for Germans, Jews, Ukrainians, and other minorities. Similar treaties were signed by Czechoslovakia, Romania, Yugoslavia, Greece, Austria, Hungary, Bulgaria, and later by Latvia, Estonia and Lithuania. Finland and Germany were not asked to sign a minority rights treaty. Interwar In the Treaty of Versailles (1919) the winners recognised the new states (Poland, Czechoslovakia, Hungary, Austria, Yugoslavia, Finland, Estonia, Latvia, Lithuania) created in central Europe from the defunct German, Austro-Hungarian and Russian empires, based on national (ethnic) self-determination. It was a peaceful era with a few small wars before 1922 such as the Ukrainian–Soviet War (1917–1921) and the Polish–Soviet War (1919–1921). Prosperity was widespread, and the major cities sponsored a youth culture called the "Roaring Twenties" or "Jazz Age" that was often featured in the cinema, which attracted very large audiences. The Allied victory in the First World War seemed to mark the triumph of liberalism, not just in the Allied countries themselves, but also in Germany and in the new states of Eastern Europe, as well as Japan. Authoritarian militarism as typified by Germany had been defeated and discredited. Historian Martin Blinkhorn argues that the liberal themes were ascendant in terms of "cultural pluralism, religious and ethnic toleration, national self-determination, free-market economics, representative and responsible government, free trade, unionism, and the peaceful settlement of international disputes through a new body, the League of Nations." However, as early as 1917, the emerging liberal order was being challenged by the new communist movement taking inspiration from the Russian Revolution. Communist revolts were beaten back everywhere else, but they did succeed in Russia. Fascism and authoritarianism Italy adopted an authoritarian dictatorship known as Fascism in 1922; it became a model for Hitler in Germany and for right wing elements in other countries. Historian Stanley G. Payne says Italian fascism was: A primarily political dictatorship....The Fascist Party itself had become almost completely bureaucratized and subservient to, not dominant over, the state itself. Big business, industry, and finance retained extensive autonomy, particularly in the early years. The armed forces also enjoyed considerable autonomy....The Fascist militia was placed under military control....The judicial system was left largely intact and relatively autonomous as well. The police continued to be directed by state officials and were not taken over by party leaders...nor was a major new police elite created....There was never any question of bringing the Church under overall subservience.... Sizable sectors of Italian cultural life retained extensive autonomy, and no major state propaganda-and-culture ministry existed....The Mussolini regime was neither especially sanguinary nor particularly repressive. Authoritarian regimes replaced democracy in the 1930s in Nazi Germany, Portugal, Austria, Poland, Greece, the Baltic countries and Francoist Spain. By 1940, there were only four liberal democracies left on the European continent: France, Finland, Switzerland and Sweden. Great Depression: 1929–1939 After the Wall Street Crash of 1929, nearly the whole world sank into a Great Depression, as money stopped flowing from New York to Europe, prices fell, profits fell, and unemployment soared. The worst hit sectors included heavy industry, export-oriented agriculture, mining and lumbering, and construction. World trade fell by two-thirds. Liberalism and democracy were discredited. In most of Europe, as well as in Japan and most of Latin America, nation after nation turned to dictators and authoritarian regimes. The most momentous change of government came when Hitler and his Nazis took power in Germany in 1933. The main institution that was meant to bring stability was the League of Nations, created in 1919. However the League failed to resolve any major crises and by 1938 it was no longer a major player. The League was undermined by the bellicosity of Nazi Germany, Imperial Japan, the Soviet Union, and Mussolini's Italy, and by the non-participation of the United States. By 1937 it was largely ignored. A major civil war took place in Spain, with the nationalists winning. The League of Nations was helpless as Italy conquered Ethiopia and Japan seized Manchuria in 1931 and took over most of China starting in 1937. The Spanish Civil War (1936–1939) was marked by numerous small battles and sieges, and many atrocities, until the rebels (the Nationalist faction), led by Francisco Franco, won in 1939. There was military intervention as Italy sent land forces, and Germany sent smaller elite air force and armoured units to the Nationalists. The Soviet Union sold armaments to the leftist Republican faction on the other side, while the Communist parties in numerous countries sent soldiers to the "International Brigades." The civil war did not escalate into a larger conflict, but did become a worldwide ideological battleground that pitted the left, the communist movement and many liberals against Catholics, conservatives, and fascists. Britain, France and the US remained neutral and refused to sell military supplies to either side. Worldwide there was a decline in pacifism and a growing sense that another world war was imminent, and that it would be worth fighting for. World War II In the Munich Agreement of 1938, Britain and France adopted a policy of appeasement as they gave Hitler what he wanted out of Czechoslovakia in the hope that it would bring peace. It did not. In 1939 Germany took over the rest of Czechoslovakia and appeasement policies gave way to hurried rearmament as Hitler next turned his attention to Poland. After allying with Japan in the Anti-Comintern Pact and then also with Benito Mussolini's Italy in the "Pact of Steel", and finally signing a non-aggression treaty with the Soviet Union in August 1939, Hitler launched the Second World War on 1 September 1939 by attacking Poland. To his surprise Britain and France declared war on Germany, but there was little fighting during the "Phoney War" period. War began in earnest in spring 1940 with the successful Blitzkrieg conquests of Denmark, Norway, the Low Countries, and France. Britain remained alone but refused to negotiate, and defeated Germany's air attacks in the Battle of Britain. Hitler's goal was to control Eastern Europe but because of his failure to defeat Britain and the Italian failures in North Africa and the Balkans, the great attack on the Soviet Union was delayed until June 1941. Despite initial successes, the Wehrmacht was stopped close to Moscow in December 1941. Over the next year the tide was turned and the Germans started to suffer a series of defeats, for example in the siege of Stalingrad and at Kursk. Meanwhile, Japan (allied to Germany and Italy since September 1940) attacked Britain and the United States on 7 December 1941; Germany then completed its over-extension by declaring war on the United States. War raged between the Axis Powers (Germany, Italy, and Japan) and the Allied Forces (British Empire, Soviet Union, and the United States). The Allied Forces won in North Africa, invaded Italy in 1943, and recaptured France in 1944. In the spring of 1945 Germany itself was invaded from the east by the Soviet Union and from the west by the other Allies. As the Red Army conquered the Reichstag in the Battle of Berlin, Hitler committed suicide and Germany surrendered in early May. World War II was the deadliest conflict in human history, causing between 50 and 80 million deaths, the majority of whom were civilians (approximately 38 to 55 million). This period was also marked by systematic genocide. In 1942–45, separately from the war-related deaths, the Nazis killed an additional number of over 11 million civilians identified through IBM-enabled censuses, including the majority of the Jews and Gypsies of Europe, millions of Polish and Soviet Slavs, and also homosexuals, Jehovah's Witnesses, misfits, disabled, and political enemies. Meanwhile, in the 1930s the Soviet system of forced labour, expulsions and allegedly engineered famine had a similar death toll. During and after the war millions of civilians were affected by forced population transfers. Cold War era The world wars ended the pre-eminent position of Britain, France and Germany in Europe and the world. At the Yalta Conference, Europe was divided into spheres of influence between the victors of World War II, and soon became the principal zone of contention in the Cold War between the two power blocs, the Western countries and the Communist bloc. The United States and the majority of European liberal democracies at the time (United Kingdom, France, Italy, Netherlands, West Germany etc.) established the NATO military alliance. Later, the Soviet Union and its satellites (Bulgaria, Czechoslovakia, East Germany, Hungary, Poland, and Romania) in 1955 established the Warsaw Pact as a counterpoint to NATO. The Warsaw Pact had a much larger ground force, but the American-French-British nuclear umbrellas protected NATO. Communist states were imposed by the Red Army in the East, while parliamentary democracy became the dominant form of government in the West. Most historians point to its success as the product of exhaustion with war and dictatorship, and the promise of continued economic prosperity. Martin Conway also adds that an important impetus came from the anti-Nazi wartime political coalitions. Economic recovery The United States gave away about $20 billion in Marshall Plan grants and other grants and low-interest long-term loans to Western Europe, 1945 to 1951. Historian Michael J. Hogan argues that American aid was critical in stabilizing the economy and politics of Western Europe. It brought in modern management that dramatically increased productivity, and encouraged cooperation between labor and management, and among the member states. Local Communist parties were opposed, and they lost prestige and influence and a role in government. In strategic terms, says Hogan, the Marshall Plan strengthened the West against the possibility of a communist invasion or political takeover. However, the Marshall Plan's role in the rapid recovery has been debated. Most reject the idea that it only miraculously revived Europe, since the evidence shows that a general recovery was already under way thanks to other aid programs from the United States. Economic historians Bradford De Long and Barry Eichengreen conclude it was, " History's Most Successful Structural Adjustment Program." They state: It was not large enough to have significantly accelerated recovery by financing investment, aiding the reconstruction of damaged infrastructure, or easing commodity bottlenecks. We argue, however, that the Marshall Plan did play a major role in setting the stage for post-World War II Western Europe's rapid growth. The conditions attached to Marshall Plan aid pushed European political economy in a direction that left its post World War II "mixed economies" with more "market" and less "controls" in the mix. The Soviet Union concentrated on its own recovery. It seized and transferred most of Germany's industrial plants and it exacted war reparations from East Germany, Hungary, Romania, and Bulgaria, using Soviet-dominated joint enterprises. It used trading arrangements deliberately designed to favor the Soviet Union. Moscow controlled the Communist parties that ruled the satellite states, and they followed orders from the Kremlin. Historian Mark Kramer concludes: The net outflow of resources from eastern Europe to the Soviet Union was approximately $15 billion to $20 billion in the first decade after World War II, an amount roughly equal to the total aid provided by the United States to western Europe under the Marshall Plan. Western Europe began economic and then political integration, with the aim to unite the region and defend it. This process included organisations such as the European Coal and Steel Community, which grew and evolved into the European Union, and the Council of Europe. The Solidarność movement in the 1980s weakened the Communist government in Poland. At the time the Soviet leader Mikhail Gorbachev initiated perestroika and glasnost, which weakened Soviet influence in Europe, particularly in the USSR. In 1989 after the Pan-European Picnic the Iron Curtain and the Berlin Wall came down and Communist governments outside the Soviet Union were deposed. In 1990 the Federal Republic of Germany absorbed East Germany, after making large cash payments to the USSR. In 1991 the Communist Party of the Soviet Union in Moscow collapsed, ending the USSR, which split into fifteen independent states. The largest, Russia, took the Soviet Union's seat on the United Nations Security Council. The most violent dissolution happened in Yugoslavia, in the Balkans. Four (Slovenia, Croatia, Bosnia and Herzegovina and North Macedonia) out of six Yugoslav republics declared independence and for most of them a violent war ensued, in some parts lasting until 1995. In 2006 Montenegro seceded and became an independent state. In the post–Cold War era, NATO and the EU have been gradually admitting most of the former members of the Warsaw Pact. Looking at the half century after the war historian Walter Lacquer concluded: "The postwar generations of European elites aimed to create more democratic societies. They wanted to reduce the extremes of wealth and poverty and provide essential social services in a way that prewar generations had not. They had had quite enough of unrest and conflict. For decades many Continental societies had more or less achieved these aims and had every reason to be proud of their progress. Europe was quiet and civilized. Europe's success was based on recent painful experience: the horrors of two world wars; the lessons of dictatorship; the experiences of fascism and communism. Above all, it was based on a feeling of European identity and common values – or so it appeared at the time." The post-war period also witnessed a significant rise in the standard of living of the Western European working class. As noted by one historical text, "within a single generation, the working classes of Western Europe came to enjoy the multiple pleasures of the consumer society." Western Europe's industrial nations in the 1970s were hit by a global economic crisis. They had obsolescent heavy industry, and suddenly had to pay very high energy prices which caused sharp inflation. Some of them also had inefficient nationalized railways and heavy industries. In the important field of computer technology, European nations lagged behind the United States. They also faced high government deficits and growing unrest led by militant labour unions. There was an urgent need for new economic directions. Germany and Sweden sought to create a social consensus behind a gradual restructuring. Germany's efforts proved highly successful. In Britain under the premiership of Margaret Thatcher, the solution was shock therapy, high interest rates, austerity, and selling off inefficient corporations as well as the public housing, which was sold off to the tenants. One result was escalating social tensions in Britain, led by the militant coal miners. Thatcher eventually defeated her opponents and radically changed the British economy, but the controversy never went away as shown by the hostile demonstrations at the time of her death in 2013. Recent history The end of the Cold War came in a series of events from 1979 to 1991, mainly in Eastern Europe. In the end, these brought the fall of the Iron Curtain, the German reunification and the end of Soviet control over their Eastern European satellites and their worldwide network of communist parties in a friendly chain reaction from the Pan-European Picnic in 1989. The finals brought the division of the Soviet Union into 15 non-communist states in 1991. Italian historian Federico Romero reports that observers at the time emphasized that: The systemic and ideological confrontation between capitalism and communism had faded away. The geopolitical partition of Europe was no more. Nuclear deterrence was morphing into a less armed, almost hypothetical version of its previous self. Superpower rivalry was rapidly wound up with cascading effects in various areas of the world. Following the end of the Cold War, the European Economic Community pushed for closer integration, co-operation in foreign and home affairs, and started to increase its membership into the neutral and former communist countries. In 1993, the Maastricht Treaty established the European Union, succeeding the EEC and furthering political co-operation. The neutral countries of Austria, Finland and Sweden acceded to the EU, and those that didn't join were tied into the EU's economic market via the European Economic Area. These countries also entered the Schengen Agreement which lifted border controls between member states. The Maastricht Treaty created a single currency for most EU members. The euro was created in 1999 and replaced all previous currencies in participating states in 2002. The most notable exception to the currency union, or eurozone, was the United Kingdom, which also did not sign the Schengen Agreement. The EU did not participate in the Yugoslav Wars, and was divided on supporting the United States in the 2003–2011 Iraq War. NATO was part of the war in Afghanistan, but at a much lower level of involvement than the United States. In 2004, the EU gained 10 new members. (Estonia, Latvia, and Lithuania, which had been part of the Soviet Union; Czech Republic, Hungary, Poland, Slovakia, and Slovenia, five former-communist countries; Malta, and the divided island of Cyprus.) These were followed by Bulgaria and Romania in 2007. Russia's regime had interpreted these expansions as violations against NATO's promise to not expand "one inch to the east" in 1990. Russia engaged in a number of bilateral disputes about gas supplies with Belarus and Ukraine which endangered gas supplies to Europe. Russia also engaged in a minor war with Georgia in 2008. Supported by the United States and some European countries, Kosovo's government unilaterally declared independence from Serbia on 17 February 2008. Public opinion in the EU turned against enlargement, partially due to what was seen as over-eager expansion including Turkey gaining candidate status. The European Constitution was rejected in France and the Netherlands, and then (as the Treaty of Lisbon) in Ireland, although a second vote passed in Ireland in 2009. The financial crisis of 2007–08 affected Europe, and government responded with austerity measures. Limited ability of the smaller EU nations (most notably Greece) to handle their debts led to social unrest, government liquidation, and financial insolvency. In May 2010, the German parliament agreed to loan 22.4 billion euros to Greece over three years, with the stipulation that Greece follow strict austerity measures. See European sovereign-debt crisis. Beginning in 2014, Ukraine has been in a state of revolution and unrest with two breakaway regions (Donetsk and Lugansk) attempting to join Russia as full federal subjects. (See Russo-Ukrainian War.) On 16 March, a disputed referendum was held in Crimea leading to the de facto secession of Crimea and its largely internationally unrecognized annexation to the Russian Federation as the Republic of Crimea. In June 2016, in a referendum in the United Kingdom on the country's membership in the European Union, 52% of voters voted to leave the EU, leading to the complex Brexit separation process and negotiations, which led to political and economic changes for both the UK and the remaining European Union countries. The UK left the EU on 31 January 2020. Later that year, Europe was affected by the COVID-19 pandemic. According to the Wall Street Journal in 2021 as Angela Merkel stepped down as the highly popular Chancellor of Germany after 16 years: Ms. Merkel leaves in her wake a weakened Europe, a region whose aspirations to act as a third superpower have come to seem ever more unrealistic. When she became chancellor in 2005, the EU was at a high point: It had adopted the euro, which was meant to rival the dollar as a global currency, and had just expanded by absorbing former members of the Soviet bloc. Today’s EU, by contrast, is geographically and economically diminished. Having lost the U.K. because of Brexit, it faces deep political and cultural divisions, lags behind in the global race for innovation and technology and is increasingly squeezed by the mounting U.S.-China strategic rivalry. Europe has endured thanks in part to Ms. Merkel’s pragmatic stewardship, but it has been battered by crises during her entire time in office.Russia began an invasion of Ukraine on 24 February 2022, in a major escalation of the Russo-Ukrainian War that began in 2014. It is the largest conventional military attack in Europe since World War II. Chronology 7000 BC: Neolithic in Europe begins. 4600 - 4200 BC: First European proto-civilisation, first golden artefacts and first fortified stone town - the Varna culture. 5000 - 3500 BC: First European proto-script - the Old European script (Danubian script). 3850 - 3600 BC: Malta's Temple period begins. 3500 BC: First European civilization, Minoan civilization, begins on Crete. 3000 BC: Indo-Europeans begin a large-scale settlement of the continent. 2500 BC: Stonehenge is constructed. 2100 BC: First European script, Cretan hieroglyphs, is invented by Minoans. 1750 BC: Mycenaean civilization begins. 1600 BC: Thera eruption occurs on the island of Santorini, destructing the Minoan city of Thera. 1450 BC: Crete is conquered by Mycenaeans. 1200 BC: Late Bronze Age collapse begins, that may be seen in the context of a technological history that saw the slow spread of ironworking technology from present-day Bulgaria and Romania in the 13th and the 12th centuries BC. 1100 BC: Minoan civilization falls. 1050 BC: Mycenaean civilization falls after a period of palace destruction, marking the beginning of Greek Dark Ages. 900 BC: Etruscan civilization begins. 800 BC: Greek Dark Ages end, marking the beginning of classical antiquity. 753 BC: Traditional year of founding of Rome. 700 BC: Homer composes The Iliad, an epic poem that represents the first extended work of European literature. 509 BC: Roman Republic is created. 499 BC: Greco-Persian Wars begin. c.480 BC: The Thracian Odrysian kingdom was founded as the most important Daco-Thracian state union. 449 BC: End of Greco-Persian Wars with Greeks defeating Achaemid Empire. 440 BC: Herodotus defends Athenian political freedom in the Histories. 404 BC: Sparta wins the Peloponnesian War. 323 BC: Alexander the Great dies and his Macedonian Empire (reaching far into Asia) fragments. 264 BC: Punic Wars begin. 146 BC: Punic Wars end with destruction of Carthage. 48 BC: Julius Caesar crosses the Rubicon river, marking the beginning of a civil war. 44 BC: Julius Caesar is murdered. The Roman Republic enters its terminal crisis. 27 BC: Establishment of the Roman Empire under Octavian. AD 14 AD: Octavian dies. 30 or 33 AD: Jesus, a popular religious leader, is crucified. 45–55 (ca): First Christian congregations in mainland Greece and in Rome. 68: First Roman imperial dynasty, Julio-Claudian, ends with suicide of Nero. 79: Eruption of Vesuvius occurs, buriing the cities of Pompeii, Herculaneum and Stabiae under the ashes. 117: Roman Empire reaches its territorial peak. 166: Antonine Plague begins. 293: Diocletian reorganizes the Empire by creating the Tetrarchy. 313: Constantine officially recognises Christianity, marking the end of the persecution of Christians. 330: Constantine makes Constantinople into his capital, a new Rome. 370: Huns first enter Europe. 395: Following the death of Theodosius I, the Empire is permanently split into the Eastern Roman Empire (later Byzantium) and the Western Roman Empire. 476: Odoacer captures Ravenna and deposes the last Roman emperor in the west: traditionally seen as the end date of the Western Roman Empire. 527: Justinian I is crowned emperor of Byzantium. Orders the editing of Corpus Juris Civilis, Digest (Roman law). 597: Beginning of Roman Catholic Christianization of Anglo-Saxon England (missions and churches had been in existence well before this date, but their contacts with Rome had been loose or nonexistent) 600: Saint Columbanus uses the term "Europe" in a letter. 655: Jus patronatus. 681: Khan Asparukh leads the Bulgars and in a union with the numerous local Slavs invades the Byzantine empire in the Battle of Ongal, creating Bulgaria. 718: Tervel of Bulgaria helps the Byzantine Empire stop the Arabic invasion of Europe, and breaks the siege of Constantinople. 722: Battle of Covadonga in the Iberian Peninsula. Pelayo, a noble Visigoth, defeats a Muslim army that tried to conquer the Cantabrian coast. This helps establish the Christian Kingdom of Asturias, and marks the beginning of the Reconquista. 732: At the Battle of Tours, the Franks stop the advance of the Arabs into Europe. 800: Coronation of Charlemagne as Holy Roman Emperor. 813: Third Council of Tours: Priests are ordered to preach in the native language of the population. 843: Treaty of Verdun. 863: Saints Cyril and Methodius arrive in Great Moravia, initiating Christian mission among the Slav peoples. 864: Boris I of Bulgaria officially baptises the whole nation, converting the non-Christian population from Tengrism, Slavic and other paganism to Christianity, and officially founding the Bulgarian Church 872: Unification of Norway. 886: Bulgarian students of Cyril and Methodius – Sava, Kliment, Naum, Gorazd, Angelariy – arrive back to Bulgaria, creating the Preslav and Ohrid Literary Schools. 893: The Cyrillic alphabet, developed during the 9th century AD at the Preslav Literary School in the First Bulgarian Empire, becomes the official Bulgarian alphabet. 895: Hungarian people led by Árpád start to settle in the Carpathian Basin. 917: In the Battle of Achelous (917) Bulgaria defeats the Byzantine empire, and Simeon I of Bulgaria is proclaimed as emperor, thus Bulgaria becomes an empire. 962: Otto I of East Francia is crowned as "Emperor" by the Pope, beginning the Holy Roman Empire. 988 Kievan Rus adopts Christianity, often seen as the origin of the Russian Orthodox Church. 1054: Start of the East–West Schism, which divides the Christian church for centuries. 1066: Successful Norman Invasion of England by William the Conqueror. 1095: Pope Urban II calls for the First Crusade. 12th century: The 12th century in literature saw an increase in the number of texts. The Renaissance of the 12th century occurs. 1128: Battle of São Mamede, formation of Portuguese sovereignty. 1131: Birth of the Kingdom of Sicily 1185: Bulgarian sovereignty was reestablished with the anti-Byzantine uprising of the Bulgarians and Vlachs 1250: Death of emperor Frederick II; end of effective ability of emperors to exercise control in Italy. 1303: The period of the Crusades is over. 1309–1378: The Avignon Papacy 1315–1317: The Great Famine of 1315–1317 in Northern Europe 1341: Petrarch, the "Father of Humanism", becomes the first poet laureate since antiquity. 1337–1453: The Hundred Years' War between England and France. 1348–1351: Black Death kills about one-third of Europe's population. 1439: Johannes Gutenberg invents first movable type and the first printing press for books, starting the Printing Revolution. 1453: Fall of Constantinople to the Ottoman Turks. 1487: The Wars of the Roses end. 1492: The Reconquista ends in the Iberian Peninsula. A Spanish expeditionary group, commanded by Christopher Columbus, lands in the New World. 1497: Vasco da Gama departs to India starting direct trade with Asia. 1498: Leonardo da Vinci paints The Last Supper in Milan as the Renaissance flourishes. 1508: Maximilian I the last ruling "King of the Romans" and the first "elected Emperor of the Romans". 1517: Martin Luther nails his 95 theses on indulgences to the door of the church in Wittenberg, triggering discussions which would soon lead to the Reformation 1519: Ferdinand Magellan and Juan Sebastián Elcano begin first global circumnavigation. Their expedition returns in 1522. 1519: Hernán Cortés begins conquest of Mexico for Spain. 1532: Francisco Pizarro begins the conquest of Peru (the Inca Empire) for Spain. 1543: Nicolaus Copernicus publishes De revolutionibus orbium coelestium (On the Revolutions of the Celestial Spheres). 1547: The Grand Duchy of Moscow becomes the Tsardom of Russia. 1582: The introduction of the Gregorian calendar; Russia refuses to adopt it until 1918. 1610: Galileo Galilei uses his telescope to discover the moons of Jupiter. 1618: The Thirty Years' War brings massive devastation to central Europe. 1648: The Peace of Westphalia ends the Thirty Years' War, and introduces the principle of the integrity of the nation state. 1687: Isaac Newton publishes Principia Mathematica, having a profound impact on The Enlightenment. 1699: Treaty of Karlowitz concludes the Austro-Ottoman War. This marks the end of Ottoman control of Central Europe and the beginning of Ottoman stagnation, establishing the Habsburg monarchy as the dominant power in Central and Southeastern Europe. 1700: Outbreak of the War of the Spanish Succession and the Great Northern War. The first would check the aspirations of Louis XIV, king of France to dominate European affairs; the second would lead to Russia's emergence as a great power and a recognizably European state. 18th century: Age of Enlightenment spurs an intellectual renaissance across Europe. 1707: The Kingdom of Great Britain is formed by the union of the Kingdom of England and the Kingdom of Scotland. 1712: Thomas Newcomen invents first practical steam engine which begins Industrial Revolution in Britain. 1721: Foundation of the Russian Empire. 1775: James Watt invents a new efficient steam engine accelerating the Industrial Revolution in Britain. 1776: Adam Smith publishes The Wealth of Nations. 1784: Immanuel Kant publishes Answering the Question: What Is Enlightenment?. 1789: Beginning of the French Revolution and end of the absolute monarchy in France. 1792–1802: French Revolutionary Wars. 1799: Napoleon comes to power, eventually consolidating his position as Emperor of the French. 1803–1815: Napoleonic Wars end in defeat of Napoleon. 1806: Napoleon abolishes the Holy Roman Empire. 1814–1815: Congress of Vienna; Treaty of Vienna; France is reduced to 1789 boundaries; Reactionary forces dominate across Europe. 1825: George Stephenson opens the Stockton and Darlington Railway the first steam train railway for passenger traffic in the world. 1830: The southern provinces secede from the United Kingdom of the Netherlands in the Belgian Revolution. 1836: Louis Daguerre invents first practical photographic method, in effect the first camera. 1838: , the first steamship built for regularly scheduled transatlantic crossings enters service. 1848: Revolutions of 1848 and publication of The Communist Manifesto. 1852: Start of the Crimean War, which ends in 1855 in a defeat for Russia. 1859: Charles Darwin publishes On the Origin of Species. 1861: Unification of Italy after victories by Giuseppe Garibaldi. 1866: First commercially successful transatlantic telegraph cable is completed. 1860s: Russia emancipates its serfs and Karl Marx completes the first volume of Das Kapital. 1870: Franco-Prussian War and the fall of the Second French Empire. 1871: Unification of Germany under the direction of Otto von Bismarck. 1873: Panic of 1873 occurs. The Long Depression begins. 1878: Re-establishment of Bulgaria, independence of Serbia, Montenegro and Romania 1885: Karl Benz invents Benz Patent-Motorwagen, the world's first automobile. 1885: First permanent citywide electrical tram system in Europe (in Sarajevo). 1895: Auguste and Louis Lumière begin exhibitions of projected films before the paying public with their cinematograph, a portable camera, printer, and projector. 1902: Guglielmo Marconi sends first transatlantic radio transmission. 1914: Archduke Franz Ferdinand of Austria is assassinated; World War I begins. 1917: Vladimir Lenin and the Bolsheviks seize power in the Russian Revolution. The ensuing Russian Civil War lasts until 1922. 1918: World War I ends with the defeat of Germany and the Central Powers. Ten million soldiers killed; collapse of Russian, German, Austrian, and Ottoman empires. 1918: Collapse of the German Empire and monarchic system; founding of Weimar Republic. 1918: Worldwide Spanish flu epidemic kills millions in Europe. 1918: Austro-Hungarian Empire dissolves. 1919: Versailles Treaty strips Germany of its colonies, several provinces and its navy and air force; limits army; Allies occupy western areas; reparations ordered. 1920: League of Nations begins operations; largely ineffective; defunct by 1939. 1921–22: Ireland divided; Irish Free State becomes independent and civil war erupts. 1922: Benito Mussolini and the Fascists take power in Italy. 1929: Worldwide Great Depression begins with stock market crash in New York City. 1933: Adolf Hitler and the Nazis take power in Germany. 1935: Italy conquers Ethiopia; League sanctions are ineffective. 1936: Start of the Spanish Civil War; ends in 1939 with victory of Nationalists who are aided by Germany and Italy. 1938: Germany escalates the persecution of Jews with Kristallnacht. 1938: Appeasement of Germany by Britain and France; Munich agreement splits Czechoslovakia; Germany seized the remainder in 1939. 1939: Britain and France hurriedly rearm; failed to arrange treaty with USSR. 1939: Adolf Hitler and Joseph Stalin agree partition of Eastern Europe in Molotov–Ribbentrop Pact. 1939: Nazi Germany invades Poland, starting the Second World War. 1940: Great Britain under Winston Churchill becomes the last nation to hold out against the Nazis after winning the Battle of Britain. 1941: U.S. begins large-scale lend-lease aid to Britain, Free France, the USSR and other Allies; Canada also provides financial aid. 1941: Germany invades the Soviet Union in Operation Barbarossa; fails to capture Moscow or Leningrad. 1942: Adolf Hitler and Nazi Germany commence the Holocaust – a Final Solution, with the murder of 6 million Jews. 1943: After Stalingrad and Kursk, Soviet forces begin recapturing Nazi-occupied territory in the East. 1944: U.S., British and Canadian armed forces invade Nazi-occupied France at Normandy. 1945: Hitler commits suicide, Mussolini is executed. World War II ends with Europe in ruins and Germany defeated. 1945: United Nations formed. 1947: The British Empire begins a process of voluntarily dismantling with the granting of independence to India and Pakistan. 1947: Cold War begins as Europe is polarized East versus West. 1948–1951: U.S. provides large sums to rebuild Western Europe through the Marshall Plan; stimulates large-scale modernization of European industries and reduction of trade restrictions. 1949: The NATO alliance is established. 1955: USSR creates a rival military coalition, the Warsaw Pact. 1950: The Schuman Declaration begins the process of European integration. 1954: The French Empire begins to be dismantled; Withdraws from Vietnam. 1956: Suez Crisis signals the end of the effective power of the British Empire. 1956: Hungarian Uprising defeated by Soviet military forces. 1957: Treaties of Rome establish the European Economic Community from 1958. 1962: The Second Vatican Council opens and begins a period of reform in the Catholic Church 1968: The May 1968 events in France lead France to the brink of revolution. 1968: The Prague Spring is defeated by Warsaw Pact military forces. The Club of Rome is founded. 1980: The Solidarność movement under Lech Wałęsa begins open, overground opposition to the Communist rule in Poland. 1985: Mikhail Gorbachev becomes leader of the Soviet Union and begins reforms which inadvertently leads to the fall of Communism and the Soviet Union. 1986: Chernobyl disaster occurs, the worst nuclear disaster in history. 1989: Communism overthrown in all the Warsaw Pact countries except the Soviet Union. Fall of the Berlin Wall (opening of unrestrained border crossings between east and west, which effectively deprived the wall of any relevance). 1990: Reunification of Germany. 1991: Breakup of Yugoslavia and the beginning of the Yugoslav Wars. 1991: Dissolution of the Soviet Union and the creation of the Commonwealth of Independent States. 1993: Maastricht Treaty establishes the European Union. 1997–99: End of European colonial empires in Asia with the handover of Hong Kong and Macau to China. 2004: Slovenia, Hungary, the Czech Republic, Slovakia, Poland, Lithuania, Latvia, Estonia, Cyprus and Malta join the European Union. 2007: Romania and Bulgaria join the European Union. 2008: The Great Recession begins. Unemployment rises in some parts of Europe. 2013: Croatia joins the European Union. 2014: Revolution of Dignity in Ukraine and the beginning of the Russo-Ukrainian War. 2015: European migrant crisis starts. 2020: The United Kingdom leaves the European Union. 2020-ongoing: COVID-19 pandemic in Europe, countries with the most cases are Russia, the United Kingdom, France, Spain, and Italy. 2022: Russian invasion of Ukraine opens with some of the most intense combat operations in Europe since the end of the Cold War. See also Prehistoric Europe Hellenistic period Roman Empire Byzantine Empire Middle Ages Renaissance Age of Discovery Major explorations after the Age of Discovery Colonialism Age of Enlightenment Atlantic Revolutions Rise of nationalism in Europe Industrial Revolution House of European History, Museum scheduled to open in 2016 in Brussels List of historians, inclusive of most major historians List of history journals#Europe List of largest European cities in history List of predecessors of sovereign states in Europe List of sovereign states by date of formation § Europe Timeline of sovereign states in Europe List of empires List of medieval great powers List of modern great powers Atlantic World History of the Mediterranean region History of Western civilization History of the Balkans Timeline of European Union history Timeline of sovereign states in Europe Genetic history of Europe History of the Romani people References Bibliography Surveys Blum, Jerome et al. The European World (2 vol. 2nd ed. 1970) university textbook; online Davies, Norman. Europe: A History (1998), advanced university textbook Gay, Peter and R.K. Webb Modern Europe: To 1815 (1973) online, university textbook Gay, Peter and R.K. Webb Modern Europe: Since 1815 (1973), university textbook Gooch, Brison D. ed. Interpreting European history. 1: From the renaissance to Napoleon (1967) online and Interpreting European history 2: from Metternich to Present (1967). online; readings by scholars McKay, John P. et al. A History of Western Society (13th ed.; 2 vol 2020) 1300 pp; university textbook Moncure, James A. ed. Research Guide to European Historical Biography: 1450–Present (4 vol 1992); 2140 pp; historiographical guide to 200 major political and military leaders Roberts, J.M. The Penguin History of Europe (1998), survey; excerpt Simms, Brendan. Europe: The Struggle for Supremacy, from 1453 to the Present (2013), survey, with emphasis on Germany Stearns, Peter, ed. The Encyclopedia of World History (2001); comprehensive global coverage Geography and atlases Cambridge Modern History Atlas (1912) online free. 141 maps Catchpole, Brian. Map History of the Modern World (1982) Darby, H. C., and H. Fullard, eds. The New Cambridge Modern History, Vol. 14: Atlas (1970) East, W. Gordon. An Historical Geography of Europe (4th ed. 1950) Haywood, John. Atlas of world history (1997) online free Horrabin, J.F. An Atlas History of the Second Great War (9 vol 1941–45) 7 vol online free Kinder, Hermann and Werner Hilgemann. Anchor Atlas of World History (2 vol. 1978); advanced analytical maps, mostly of Europe O'Brian, Patrick K. Atlas of World History (2007) Online free Rand McNally Atlas of World History (1983), maps #76-81. Published in Britain as the Hamlyn Historical Atlas online free Robertson, Charles Grant. An historical atlas of modern Europe from 1789 to 1922 with an historical and explanatory text (1922) online free Schenk, Frithjof Benjamin, Mental Maps: The Cognitive Mapping of the Continent as an Object of Research of European History, EGO – European History Online, Mainz: Institute of European History, 2013, retrieved: 4 March 2020 Talbert, Richard J.A. Barrington Atlas of the Greek and Roman World for iPad (Princeton UP 2014) ; 102 interactive color maps from archaic Greece to the Late Roman Empire. Historical Atlas Wikipedia maps; no copyright Atlas of Germany Wikipedia maps; no copyright Major nations Carr, Raymond, ed. Spain: A history (2000) Clark, Christopher M. Iron kingdom: the rise and downfall of Prussia, 1600-1947 (2006) Davies, Norman. The Isles: A History (2001), Britain and Ireland Fraser, Rebecca. The Story of Britain: From the Romans to the Present: A Narrative History (2006) Holborn, Hajo. vol 1: A History of Modern Germany: The Reformation; vol 2: A History of Modern Germany: 1648-1840; vol 3: A history of modern Germany: 1840–1945 (1959). a standard scholarly survey. Kamen, Henry. A concise history of Spain (1973) Holmes, George, ed. The Oxford illustrated history of medieval Europe (2001). Holmes, George, ed. The Oxford illustrated history of Italy (1997) Jones, Colin. The Cambridge Illustrated History of France (1999) Kitchen, Martin The Cambridge Illustrated History of Germany (1996). Morgan, Kenneth O., ed. The Oxford illustrated history of Britain (1984) Riasanovsky, Nicholas V., and Mark D. Steinberg. A History of Russia (2 vol. 2010) Sagarra, Eda. A social history of Germany (2003) Tombs, Robert, The English and their History (2014) advanced history; online review Classical Boardman, John, et al. eds. The Oxford History of Greece and the Hellenistic World (2nd ed. 2002) 520 pp Boardman, John, et al. eds. The Oxford History of the Roman World (2001) Cartledge, Paul. The Cambridge Illustrated History of Ancient Greece (2002) Late Roman Heather, Peter. Empires and Barbarians: The Fall of Rome and the Birth of Europe (Oxford University Press; 2010); 734 pages; Examines the migrations, trade, and other phenomena that shaped a recognizable entity of Europe in the first millennium. Jones, A.H.M. The Later Roman Empire, 284–602: A Social, Economic, and Administrative Survey (2 Vol. 1964) Mitchell, Stephen. A History of the Later Roman Empire, AD 284–641: The Transformation of the Ancient World (2006) Medieval Davis, R.H.C. A History of Medieval Europe (2nd ed. 2000) Ferguson, Wallace K. Europe in Transition, 1300–1520 (1962) Hanawalt, Barbara. The Middle Ages: An Illustrated History (1999) Holmes, George, ed. The Oxford Illustrated History of Medieval Europe (2001) Koenigsberger, H.G. Medieval Europe 400–1500 (1987) Riddle, John M. A history of the Middle Ages, 300–1500 (2008) Early modern Cameron, Euan. Early Modern Europe: An Oxford History (2001) Friedrich, Carl J. The Age of the Baroque, 1610-1660 (1962); Despite the title, this is a wide-ranging Social, cultural, political and diplomatic history of Europe; 14-day borrowing copy Hesmyr, Atle. Scandinavia in the Early Modern Era; From Peasant Revolts and Witch Hunts to Constitution Drafting Yeomen (2014) Rice, Eugene F. The Foundations of Early Modern Europe, 1460-1559 (2nd ed. 1994) 240 pp Merriman, John. A History of Modern Europe: From the Renaissance to the Present (3rd ed. 2010, 2 vol), 1412 pp online Scott, Hamish, ed. The Oxford Handbook of Early Modern European History, 1350–1750: Volume I: Peoples and Place (2015). Stoye, John. Europe Unfolding, 1648-1688 (2nd ed. 2000). Wiesner, Merry E. Early Modern Europe, 1450-1789 (Cambridge History of Europe) (2006) 19th century Anderson, M.S. The Ascendancy of Europe: 1815-1914 (3rd ed. 2003) Berger, Stefan, ed. A Companion to Nineteenth-Century Europe, 1789–1914 (2006), 545pp; emphasis on historiography of 32 topics; online Blanning, T.C.W. ed. The Nineteenth Century: Europe 1789–1914 (Short Oxford History of Europe) (2000) 320 pp Brinton, Crane. A Decade of Revolution, 1789–1799 (1934) online in Langer series on history of Europe. Bruun, Geoffrey. Europe and the French Imperium, 1799–1814 (1938) online. Cameron, Rondo. France and the Economic Development of Europe, 1800–1914: Conquests of Peace and Seeds of War (1961), wide-ranging economic and business history. online Evans, Richard J. The Pursuit of Power: Europe 1815–1914 (2016), 934 pp; scholarly survey Gildea, Robert. Barricades and Borders: Europe 1800–1914 (Short Oxford History of the Modern World) (3rd ed. 2003) 544 p Gooch, Brison D. Europe in the nineteenth century; a history (1970) online Kertesz, G.A. ed Documents in the Political History of the European Continent 1815-1939 (1968), 507 pp; several hundred short documents; primary sources Mason, David S. A Concise History of Modern Europe: Liberty, Equality, Solidarity (2011), since 1700 Merriman, John, and J.M. Winter, eds. Europe 1789 to 1914: Encyclopedia of the Age of Industry and Empire (5 vol. 2006) Ramm, Agatha. Europe in the Nineteenth Century 1789–1905 (1984). 1950 edition online Richardson, Hubert N.B. A Dictionary of Napoleon and His Times (1921) online free 489 pp Steinberg, Jonathan. Bismarck: A Life (2011) Salmi, Hannu. 19th Century Europe: A Cultural History (2008). Taylor, A.J.P. The Struggle for Mastery in Europe 1848-1918 (1954) online free; Advanced diplomatic history online Thomson, David. Europe Since Napoleon (1923) 524 pp online Since 1900 Brose, Eric Dorn. A History of Europe in the Twentieth Century (2004) 548 pp Brown, Archie. The Rise and Fall of Communism (2009), global Buchanan, Tom. Europe's Troubled Peace: 1945 to the Present (Blackwell History of Europe) (2012) Cook, Bernard A. Europe Since 1945: An Encyclopedia (2 vol; 2001), 1465 pp Davies, Norman. Europe at War 1939-1945: No Simple Victory (2008) Dear, I.C.B. and M.R.D. Foot, eds. The Oxford Companion to World War II (2006) Frank, Matthew. Making Minorities History: Population Transfer in Twentieth-Century Europe (Oxford UP, 2017). 464 pp. online review Grenville, J.A.S. A History of the World in the Twentieth Century (1994). online free Hallock, Stephanie A. The World in the 20th Century: A Thematic Approach (2012) Judt, Tony. Postwar: A History of Europe Since 1945 (2006) Martel, Gordon, ed. A Companion to Europe, 1900-1950 (2011) 32 essays by scholars; emphasis on historiography Mazower, Mark. Dark Continent: Europe's Twentieth Century (2000) 512 pp Merriman, John, and Jay Winter, eds. Europe Since 1914: Encyclopedia of the Age War and Reconstruction (5 vol. 2006) Paxton, Robert O., and Julie Hessler. Europe in the twentieth century (5th edition 2011. Pollard, Sidney, ed. Wealth and Poverty: an Economic History of the 20th Century (1990), 260 pp; global perspective online free Sontag, Raymond James. A broken world, 1919–1939 (1972), wide-ranging history of interwar Europe; online free to borrow Stone, Dan, ed. The Oxford Handbook of Postwar European History (2015). Ther, Philipp. Europe since 1989: A History (Princeton UP, 2016) excerpt, 440 pp Toynbee, Arnold, ed. Survey Of International Affairs: Hitler's Europe 1939–1946 (1954) online Wasserstein, Bernard. Barbarism and civilization: A history of Europe in our time (2007), since 1914 online , on World War II Agriculture and economy Berend, Iván T. An Economic History of Twentieth-Century Europe (2013) Gras, Norman. A history of agriculture in Europe and America (1925). online Kander, Astrid, et al. Power to the People: Energy in Europe Over the Last Five Centuries (Princeton UP, 2013) Milward, Alan S. and S.B. Saul. The Development of the Economies of Continental Europe: 1850–1914 (1977) online Murray, Jacqueline. The First European Agriculture (1970) Pollard, Sidney, ed. Wealth and Poverty: an Economic History of the 20th Century (1990), 260 pp; global perspective online free Pounds, N.J.G. An Economic History of Medieval Europe (1994) Slicher van Bath, B.H. The agrarian history of Western Europe, AD 500–1850 (1966) Thorp, William Long. Business Annals: United States, England, France, Germany, Austria, Russia, Sweden, Netherlands, Italy, Argentina, Brazil, Canada, South Africa, Australia, India, Japan, China (1926) capsule summary of conditions in each country for each quarter-year 1790–1925 online Diplomacy Albrecht-Carrié, René. A Diplomatic History of Europe Since the Congress of Vienna (1958), 736pp; a basic introduction, 1815–1955 online Black, Jeremy. A History of Diplomacy (2011) Kertesz, G.A. ed Documents in the Political History of the European Continent 1815-1939 (1968), 507 pp; several hundred short documents Langer, William. An Encyclopedia of World History (5th ed. 1973), very detailed outline Macmillan, Margaret. The War That Ended Peace: The Road to 1914 (2013) cover 1890s to 1914 Mowat, R.B. History of European Diplomacy, 1451–1789 (1928) 324 pages Petrie, Charles. Earlier diplomatic history, 1492–1713 (1949), covers all of Europe; online Petrie, Charles. Diplomatic History, 1713–1933 (1946), broad summary online Schroeder, Paul. The Transformation of European Politics 1763–1848 (1994) online; advanced diplomatic history Sontag, Raymond James. European Diplomatic History 1871–1932 (1933) online Steiner, Zara. The Lights that Failed: European International History 1919–1933 (2007) Steiner, Zara. The Triumph of the Dark: European International History 1933–1939 (2011) Taylor, A.J.P The struggle for mastery in Europe, 1848–1918 (1954) online Empires and interactions Bayly, C.A. ed. Atlas of the British Empire (1989). survey by scholars; heavily illustrated Brendon, Piers. The Decline and Fall of the British Empire, 1781–1997 (2008), wide-ranging survey Cotterell, Arthur. Western Power in Asia: Its Slow Rise and Swift Fall, 1415 – 1999 (2009) popular history; excerpt Darwin, John. After Tamerlane: The Rise and Fall of Global Empires, 1400–2000 (2008). James, Lawrence. The Rise and Fall of the British Empire (1997) Poddar, Prem, and Lars Jensen, eds., A historical companion to postcolonial literatures: Continental Europe and Its Empires (Edinburgh UP, 2008), excerpt also entire text online Ideas and science Heilbron, John L., ed. The Oxford Companion to the History of Modern Science (2003) Wiener, Philip P. Dictionary of the History of Ideas (5 vol 1973) Religion Forlenza, Rosario. "New Perspectives on Twentieth-Century Catholicism." Contemporary European History 28.4 (2019): 581–595 DOI: https://doi.org/10.1017/S0960777319000146 Latourette, Kenneth Scott. Christianity in a Revolutionary Age: A History of Christianity in the Nineteenth and Twentieth Centuries (5 vol. 1958–69) vol 1, 2, and 4 for detailed country-by-country coverage MacCulloch, Diarmaid. Christianity: The First Three Thousand Years (2011) Social Maynes, Mary Jo. Schooling in Western Europe: A social history (1985). Patel, Klaus Kiran, Transnational History, EGO – European History Online, Mainz: Institute of European History, 2010, retrieved: 4 March 2020 Schmale, Wolfgang, A Transcultural History of Europe – Perspectives from the History of Migration, EGO – European History Online, Mainz: Institute of European History, 2010, retrieved: 4 March 2020 Stearns, Peter N., ed. Encyclopedia of European Social History (6 vol 2000), 3000 pp comprehensive coverage Stearns, Peter N., and Herrick Chapman. European society in upheaval: social history since 1750 (1975). Tipton, F. and Robert Aldrich. An Economic and Social History of Europe (1987), two volumes Warfare Archer, Christon I.; John R. Ferris, Holger H. Herwig. World History of Warfare (2002) The Cambridge History of the First World War (3 vol 2014) online The Cambridge History of the Second World War (3 vol 2015) online Cruttwell, C. R. M. F. A History of the Great War, 1914–1918 (1934), general military history online free Dupuy, R. Ernest, The Harper Encyclopedia of Military History: From 3500 BC to the Present (1993) Gerwarth, Robert, and Erez Manela. "The Great War as a Global War: Imperial Conflict and the Reconfiguration of World Order, 1911–1923." Diplomatic History 38.4 (2014): 786–800. online Gerwarth, Robert, and Erez Manela, eds. Empires at War: 1911–1923 (2014), 12 very wide-ranging essays by scholars. excerpt Goldsworthy, Adrian, and John Keegan. Roman Warfare (2000) Horne, John, ed. A Companion to World War I (2012) Keegan, John. A History of Warfare (1994) online Kennedy, Paul. The Rise and Fall of the Great Powers (1989) online Muehlbauer, Matthew S., and David J. Ulbrich, eds. The Routledge History of Global War and Society (Routledge, 2018) Paret, Peter, ed. Makers of Modern Strategy (1986), ideas of warfare Sharman, Jason C. "Myths of military revolution: European expansion and Eurocentrism." European Journal of International Relations 24.3 (2018): 491-513 online Stevenson, David. Cataclysm: The First World War As Political Tragedy (2004) major reinterpretation, 560pp Strachan, Hew. The First World War: Volume I: To Arms (2004): a major scholarly synthesis. Thorough coverage of 1914–16; 1245pp ; comprehensive overview with emphasis on diplomacy Winter, J. M. The Experience of World War I (2nd ed 2005), topical essays; Winter, Jay, and Antoine Prost (2nd ed 2020). The Great War in History: Debates and Controversies, 1914 to the Present. New York: Cambridge University Press. Winter, Jay. Remembering War: The Great War Between Memory and History in the Twentieth Century. (Yale University Press, 2006) Winter, Jay, ed. (2014). The Cambridge History of the First World War (2 vol. Cambridge University Press, 2014) Zeiler, Thomas W. and Daniel M. DuBois, eds. A Companion to World War II (2 vol 2013), 1030 pp; comprehensive overview by scholars Women and gender Anderson, Bonnie S. and Judith P. Zinsser. A History of Their Own: Women in Europe from Prehistory to the Present (2nd ed 2000) Bridenthal, Renate, et al. eds. Becoming Visible: Women in European History (3rd ed. 1997), 608 pp; essays by scholars Frey, Linda, Marsha Frey, Joanne Schneider. Women in Western European History: A Select Chronological, Geographical, and Topical Bibliography (1982) Hufton, Olwen. The Prospect Before Her: A History of Women in Western Europe, 1500–1800'' (1996) External links EurhistXX: The Network for the Contemporary History of Europe, edited in English from Berlin Contains information on historical trends in living standards in various European countries European History Primary Sources Online access to primary sources for historians Vistorica – Timelines of European modern history
13432
https://en.wikipedia.org/wiki/Foreign%20relations%20of%20Hungary
Foreign relations of Hungary
Hungary wields considerable influence in Central and Eastern Europe and is a middle power in international affairs. The foreign policy of Hungary is based on four basic commitments: to Atlantic co-operation, to European integration, to international development and to international law. The Hungarian economy is fairly open and relies strongly on international trade. Hungary has been a member of the United Nations since December 1955 and holds current membership with the European Union, NATO, the OECD, the Visegrád Group, the WTO, the World Bank, the AIIB and the IMF. Hungary took on the presidency of the Council of the European Union for half a year in 2011 and the next will be in 2024. In 2015, Hungary was the fifth largest OECD Non-DAC donor of development aid in the world, which represents 0.13% of its Gross National Income. In this regard, Hungary stands before Spain, Israel or Russia. Hungary's capital city, Budapest is home to more than 100 embassies and representative bodies as an international political actor. Hungary hosts the main and regional headquarters of many international organizations as well, including European Institute of Innovation and Technology, European Police College, United Nations High Commissioner for Refugees, Food and Agriculture Organization of the United Nations, International Centre for Democratic Transition, Institute of International Education, International Labour Organization, International Organization for Migration, International Red Cross, Regional Environmental Center for Central and Eastern Europe, Danube Commission and even others. Since 1989, Hungary's top foreign policy goal was achieving integration into Western economic and security organizations. Hungary joined the Partnership for Peace program in 1994 and has actively supported the IFOR and SFOR missions in Bosnia. Hungary since 1989 has also improved its often frosty neighborly relations by signing basic treaties with Ukraine, Slovakia, and Romania. These renounce all outstanding territorial claims and lay the foundation for constructive relations. However, the issue of ethnic Hungarian minority rights in Romania, Slovakia and Ukraine periodically causes bilateral tensions to flare up. Hungary since 1989 has signed all of the OSCE documents, and served as the OSCE's Chairman-in-Office in 1997. Hungary's record of implementing CSCE Helsinki Final Act provisions, including those on the reunification of divided families, remains among the best in Central and Eastern Europe. Except for the short-lived neutrality declared by the anti-Soviet leader Imre Nagy in November 1956, Hungary's foreign policy generally followed the Soviet lead from 1947 to 1989. During the Communist period, Hungary maintained treaties of friendship, cooperation, and mutual assistance with the Soviet Union, Poland, Czechoslovakia, the German Democratic Republic, Romania, and Bulgaria. It was one of the founding members of the Soviet-led Warsaw Pact and Comecon, and it was the first country to withdraw from those organizations. After 1989, Hungary oriented more towards the West, joined NATO in 1999 and the European Union in 2004. Overview As with any country, Hungarian security attitudes are shaped largely by history and geography. For Hungary, this is a history of more than 400 years of domination by great powers—the Ottomans, the Habsburg dynasty, the Germans during World War II, and the Soviets during the Cold War—and a geography of regional instability and separation from Hungarian minorities living in neighboring countries. Hungary's foreign policy priorities, largely consistent since 1990, represent a direct response to these factors. Since 1990, Hungary's top foreign policy goal has been achieving integration into Western economic and security organizations. Hungary joined the Partnership for Peace program in 1994 and has actively supported the IFOR and SFOR missions in Bosnia. The Horn government achieved Hungary's most important foreign policy successes of the post-communist era by securing invitations to join both NATO and the European Union in 1997. Hungary became a member of NATO in 1999, and a member of the EU in 2004. Hungary also has improved its often frosty neighborly relations by signing basic treaties with Romania, Slovakia, and Ukraine. These renounce all outstanding territorial claims and lay the foundation for constructive relations. However, the issue of ethnic Hungarian minority rights in Slovakia and Romania periodically causes bilateral tensions to flare up. Hungary was a signatory to the Helsinki Final Act in 1975, has signed all of the CSCE/OSCE follow-on documents since 1989, and served as the OSCE's Chairman-in-Office in 1997. Hungary's record of implementing CSCE Helsinki Final Act provisions, including those on the reunification of divided families, remains among the best in eastern Europe. Hungary has been a member of the United Nations since December 1955. The Gabčíkovo - Nagymaros Dams project This involves Hungary and Czechoslovakia, and was agreed on September 16, 1977 ("Budapest Treaty"). The treaty envisioned a cross-border barrage system between the towns Gabčíkovo, Czechoslovakia and Nagymaros, Hungary. After an intensive campaign, the project became widely hated as a symbol of the old communist regime. In 1989 the Hungarian government decided to suspend it. In its sentence from September 1997, the International Court of Justice stated that both sides breached their obligation and that the 1977 Budapest Treaty is still valid. In 1998 the Slovak government turned to the International Court, demanding the Nagymaros part to be built. The international dispute was partially solved in 2017. On March 19, 2008 Hungary recognized Kosovo as an independent country. Illicit drugs: Major trans-shipment point for Southwest Asian heroin and cannabis and transit point for South American cocaine destined for Western Europe; limited producer of precursor chemicals, particularly for amphetamines and methamphetamines Refugee protection: The Hungarian border barrier was built in 2015, and Hungary was criticized by other European countries for using tear gas and water cannons on refugees of the Syrian Civil War as they were trying to pass the country. Since 2017, Hungary–Ukraine relations have rapidly deteriorated over the issue of the Hungarian minority in Ukraine. Hungary and Central Asia A number of Hungarian anthropologists and linguists have long had an interest in the Turkic peoples, fueled by the eastern origin of the Hungarians' ancestors. The Hungarian ethnomusicologist Bence Szabolcsi explained this motivation as follows: "Hungarians are the outermost branch leaning this way from age-old tree of the great Asian musical culture rooted in the souls of a variety of peoples living from China through Central Asia to the Black Sea". Diplomatic relations List of countries with which Hungary maintains diplomatic relations: Hungary maintains diplomatic relations with all UN member states excluding Bhutan and Vanuatu. Relations by region and country Africa Americas Asia Europe Oceania Foreign criticism In December 2010, the Fidesz government adopted a press and media law which threatens fines on media that engage in "unbalanced coverage". The law aroused criticism in the European Union as possibly "a direct threat to democracy". In 2013, the government adopted a new constitution that modified several aspects of the institutional and legal framework in Hungary. These changes have been criticized by the Council of Europe, the European Union and Human Rights Watch as possibly undermining the rule of law and human rights protection. See also List of diplomatic missions in Hungary List of diplomatic missions of Hungary Visa requirements for Hungarian citizens References Further reading Borhi, László, “In the Power Arena: U.S.-Hungarian Relations, 1942–1989,” The Hungarian Quarterly (Budapest), 51 (Summer 2010), pp 67–81. Glant, Tibor, “Ninety Years of United States-Hungarian Relations,” Eger Journal of American Studies, 13 (2012), pp 163–83. Hornyak, Arpad. Hungarian-Yugoslav Diplomatic Relations, 1918–1927 (East European Monographs, distributed by Columbia University Press; 2013) 426 pages. Niklasson, Tomas. "Regime stability and foreign policy change: interaction between domestic and foreign policy in Hungary 1956-1994" (PhD dissertation Lund University, 2006) online. Váli, Ferenc A, "The Foreign Policy of Hungary" in Kuhlman, James A (ed.), The Foreign Policies of Eastern Europe: Domestic and International Determinants (Sijthoff, Leyden, 1978).
13606
https://en.wikipedia.org/wiki/Half-life
Half-life
Half-life (symbol ) is the time required for a quantity (of substance) to reduce to half of its initial value. The term is commonly used in nuclear physics to describe how quickly unstable atoms undergo radioactive decay or how long stable atoms survive. The term is also used more generally to characterize any type of exponential (or, rarely, non-exponential) decay. For example, the medical sciences refer to the biological half-life of drugs and other chemicals in the human body. The converse of half-life (in exponential growth) is doubling time. The original term, half-life period, dating to Ernest Rutherford's discovery of the principle in 1907, was shortened to half-life in the early 1950s. Rutherford applied the principle of a radioactive element's half-life in studies of age determination of rocks by measuring the decay period of radium to lead-206. Half-life is constant over the lifetime of an exponentially decaying quantity, and it is a characteristic unit for the exponential decay equation. The accompanying table shows the reduction of a quantity as a function of the number of half-lives elapsed. Probabilistic nature A half-life often describes the decay of discrete entities, such as radioactive atoms. In that case, it does not work to use the definition that states "half-life is the time required for exactly half of the entities to decay". For example, if there is just one radioactive atom, and its half-life is one second, there will not be "half of an atom" left after one second. Instead, the half-life is defined in terms of probability: "Half-life is the time required for exactly half of the entities to decay on average". In other words, the probability of a radioactive atom decaying within its half-life is 50%. For example, the image on the right is a simulation of many identical atoms undergoing radioactive decay. Note that after one half-life there are not exactly one-half of the atoms remaining, only approximately, because of the random variation in the process. Nevertheless, when there are many identical atoms decaying (right boxes), the law of large numbers suggests that it is a very good approximation to say that half of the atoms remain after one half-life. Various simple exercises can demonstrate probabilistic decay, for example involving flipping coins or running a statistical computer program. Formulas for half-life in exponential decay An exponential decay can be described by any of the following four equivalent formulas: where is the initial quantity of the substance that will decay (this quantity may be measured in grams, moles, number of atoms, etc.), is the quantity that still remains and has not yet decayed after a time , is the half-life of the decaying quantity, is a positive number called the mean lifetime of the decaying quantity, is a positive number called the decay constant of the decaying quantity. The three parameters , , and are directly related in the following way:where is the natural logarithm of 2 (approximately 0.693). Half-life and reaction orders In chemical kinetics, the value of the half-life depends on the reaction order: Decay by two or more processes Some quantities decay by two exponential-decay processes simultaneously. In this case, the actual half-life can be related to the half-lives and that the quantity would have if each of the decay processes acted in isolation: For three or more processes, the analogous formula is: For a proof of these formulas, see Exponential decay § Decay by two or more processes. Examples There is a half-life describing any exponential-decay process. For example: As noted above, in radioactive decay the half-life is the length of time after which there is a 50% chance that an atom will have undergone nuclear decay. It varies depending on the atom type and isotope, and is usually determined experimentally. See List of nuclides. The current flowing through an RC circuit or RL circuit decays with a half-life of or , respectively. For this example the term half time tends to be used rather than "half-life", but they mean the same thing. In a chemical reaction, the half-life of a species is the time it takes for the concentration of that substance to fall to half of its initial value. In a first-order reaction the half-life of the reactant is , where (also denoted as ) is the reaction rate constant. In non-exponential decay The term "half-life" is almost exclusively used for decay processes that are exponential (such as radioactive decay or the other examples above), or approximately exponential (such as biological half-life discussed below). In a decay process that is not even close to exponential, the half-life will change dramatically while the decay is happening. In this situation it is generally uncommon to talk about half-life in the first place, but sometimes people will describe the decay in terms of its "first half-life", "second half-life", etc., where the first half-life is defined as the time required for decay from the initial value to 50%, the second half-life is from 50% to 25%, and so on. In biology and pharmacology A biological half-life or elimination half-life is the time it takes for a substance (drug, radioactive nuclide, or other) to lose one-half of its pharmacologic, physiologic, or radiological activity. In a medical context, the half-life may also describe the time that it takes for the concentration of a substance in blood plasma to reach one-half of its steady-state value (the "plasma half-life"). The relationship between the biological and plasma half-lives of a substance can be complex, due to factors including accumulation in tissues, active metabolites, and receptor interactions. While a radioactive isotope decays almost perfectly according to so-called "first order kinetics" where the rate constant is a fixed number, the elimination of a substance from a living organism usually follows more complex chemical kinetics. For example, the biological half-life of water in a human being is about 9 to 10 days, though this can be altered by behavior and other conditions. The biological half-life of caesium in human beings is between one and four months. The concept of a half-life has also been utilized for pesticides in plants, and certain authors maintain that pesticide risk and impact assessment models rely on and are sensitive to information describing dissipation from plants. In epidemiology, the concept of half-life can refer to the length of time for the number of incident cases in a disease outbreak to drop by half, particularly if the dynamics of the outbreak can be modeled exponentially. See also Half time (physics) List of radioactive nuclides by half-life Mean lifetime Median lethal dose References External links https://www.calculator.net/half-life-calculator.html Comprehensive half-life calculator Welcome to Nucleonica, Nucleonica.net (archived 2017) wiki: Decay Engine, Nucleonica.net (archived 2016) System Dynamics – Time Constants, Bucknell.edu Researchers Nikhef and UvA measure slowest radioactive decay ever: Xe-124 with 18 billion trillion years https://academo.org/demos/radioactive-decay-simulator/ Interactive radioactive decay simulator demonstrating how half-life is related to the rate of decay Chemical kinetics Radioactivity Nuclear fission Temporal exponentials
13773
https://en.wikipedia.org/wiki/Hrad%C4%8Dany
Hradčany
Hradčany (; ), the Castle District, is the district of the city of Prague, Czech Republic surrounding Prague Castle. The castle is one of the biggest in the world at about in length and an average of about wide. Its history stretches back to the 9th century. St Vitus Cathedral is located in the castle area. Most of the district consists of noble historical palaces. There are many other attractions for visitors: romantic nooks, peaceful places and beautiful lookouts. Hradčany was an independent borough until 1784, when the four independent boroughs that had formerly constituted Prague were proclaimed a single city. The other three were Malá Strana (, Lesser Quarter), Staré Město (, Old Town) and Nové Město (, New Town). References External links Official Website of the City of Prague Hradčany - Prague-wiki Photo gallery Districts of Prague
13920
https://en.wikipedia.org/wiki/Homeland
Homeland
A homeland is a place where a cultural, national, or racial identity has formed. The definition can also mean simply one's country of birth. When used as a proper noun, the Homeland, as well as its equivalents in other languages, often has ethnic nationalist connotations. A homeland may also be referred to as a fatherland, a motherland, or a mother country, depending on the culture and language of the nationality in question. Motherland Motherland refers to a mother country, i.e. the place in which somebody grew up or had lived for a long enough period that somebody has formed their own cultural identity, the place that one's ancestors lived for generations, or the place that somebody regards as home, or a Metropole in contrast to its colonies. People often refer to Mother Russia as a personification of the Russian nation. The Philippines is also considered as a motherland which is derived from the word "Inang Bayan" which means "Motherland". Within the British Empire, many natives in the colonies came to think of Britain as the mother country of one, large nation. India is often personified as Bharat Mata (Mother India). The French commonly refer to France as "la mère patrie"; Hispanic countries that were former Spanish colonies commonly referred to Spain as "la Madre Patria". Romans and the subjects of Rome saw Italy as the motherland (patria or terrarum parens) of the Roman Empire, in contrast to Roman provinces. Turks refer to Turkey as "ana vatan" (lit: mother homeland.) Fatherland Fatherland is the nation of one's "fathers", "forefathers" or ancestors. The word can also mean the country of nationality, the country in which somebody grew up, the country that somebody's ancestors lived in for generations, or the country that somebody regards as home, depending on how the individual uses it. It can be viewed as a nationalist concept, in so far as it is evocative of emotions related to family ties and links them to national identity and patriotism. It can be compared to motherland and homeland, and some languages will use more than one of these terms. The national anthem of the Netherlands between 1815 and 1932, "Wien Neêrlands Bloed", makes extensive use of the parallel Dutch word, as does the current Dutch national anthem, Het Wilhelmus. The Ancient Greek patris, fatherland, led to patrios, of our fathers and thence to the Latin patriota and Old French patriote, meaning compatriot; from these the English word patriotism is derived. The related Ancient Roman word Patria led to similar forms in modern Romance languages. "Fatherland" was first encountered by the vast majority of citizens in countries that did not themselves use it during World War II, when it was featured in news reports associated with Nazi Germany. German government propaganda used its appeal to nationalism when making references to Germany and the state. It was used in Mein Kampf, and on a sign in a German concentration camp, also signed, Adolf Hitler. The term fatherland (Vaterland) is used throughout German-speaking Europe, as well as in Dutch. National history is usually called vaderlandse geschiedenis in Dutch. Another use of the Dutch word is well known from the national anthem, "Het Wilhelmus". In German, the word became more prominent in the 19th century. It appears in numerous patriotic songs and poems, such as Hoffmann's song Lied der Deutschen which became the national anthem in 1922. Because of the use of Vaterland in Nazi-German war propaganda, the term "Fatherland" in English has become associated with domestic British and American anti-Nazi propaganda during World War II. This is not the case in Germany itself, where the word remains used in the usual patriotic contexts. Terms equating "Fatherland" in other Germanic languages: Afrikaans: Vaderland Danish: fædreland Dutch: vaderland (as in the national anthem Wilhelmus) West Frisian: heitelân German: Vaterland (as in the national anthem Das Lied der Deutschen) Icelandic: föðurland Norwegian: fedreland Scots: Swedish: fäderneslandet (besides the more common fosterlandet) A corresponding term is often used in Slavic languages, in: Russian otechestvo (отечество) or otchizna (отчизна) Polish ojczyzna in common language literally meaning "fatherland", ziemia ojców literally meaning "land of fathers", sometimes used in the phrase ziemia ojców naszych literally meaning "land of our fathers" (besides rarer name macierz "motherland") Ukrainian batʹkivshchyna (батьківщина) or vitchyzna (вітчизна). Czech otčina (although the normal Czech term for "homeland" is vlast) the Belarusians as (Baćkaŭščyna) Serbo-Croatian otadžbina (отаџбина) meaning "fatherland", domovina (домовина) meaning "homeland", dedovina (дедовина) or djedovina meaning "grandfatherland" or "land of grandfathers" Bulgarian татковина (tatkovina) as well as otechestvo (Отечество) Macedonian татковина (tatkovina) Other groups that refer to their native country as a "fatherland" Groups with languages that refer to their native country as a "fatherland" include: the Arabs as 'arḍ al-'abā' ("land of the fathers") the Armenians as (Hayreniq) the Albanians as Atdhe the Amharas as (Abbat Ager) the Austrians as Vaterland the Arakaneses as (အဖရခိုင်ပြည်) the Azerbaijanis as vətən (from Arabic) the Chechens as "Daimokh" the Estonians as isamaa (as in the national anthem Mu isamaa, mu õnn ja rõõm) the Finns as isänmaa the French, as La patrie the Flemings as Vaderland the Georgians as Samshoblo (სამშობლო - "[land] of parents") or Mamuli (მამული) the Ancient Greeks as πατρίς patris the Greeks as πατρίδα the Irish as Athartha the Kazakhs as atameken the Kyrgyz as ata meken the Latvians as tēvzeme the Liechtensteiners as Vaterland the Lithuanians as tėvynė the Nigerians as fatherland the Oromo as Biyya Abaa the Pakistanis as Vatan (madar-e-watan means motherland. Not fatherland) the Somali as Dhulka Abaa, land of the father the Swiss as Vaterland (as in the national anthem Swiss Psalm) the Thais as pituphum (ปิตุภูมิ), the word is adapted from Sanskrit the Tibetans as (pha yul) the Welsh as , 'the ancient land of my fathers' Romance languages In Romance languages, a common way to refer to one's home country is Patria/Pátria/Patrie which has the same connotation as Fatherland, that is, the nation of our parents/fathers (From the Latin, Pater, father). As patria has feminine gender, it is usually used in expressions related to one's mother, as in Italian la Madrepatria, Spanish la Madre Patria or Portuguese a Pátria Mãe (Mother Fatherland). Examples include: the Esperantists as patrio, patrolando or patrujo Aragonese, Asturian, Franco-Provençal, Galician, Italian, Spanish (in its many dialects): Patria Catalan: Pàtria Occitans: Patrìo French: Patrie Romanian: Patrie Portuguese: Pátria Multiple references to parental forms the Armenians, as Hayrenik (Հայրենիք), home. The national anthem Mer Hayrenik translates as Our Fatherland the Azerbaijanis as Ana vətən (lit. mother homeland) or Ata ocağı (lit. father's hearth) the Bosniaks as Otadžbina (Отаџбина), although Domovina (Домовина) is sometimes used colloquially meaning homeland the Chinese as zǔguó (祖国 or 祖國 (traditional chinese), "land of ancestors"), zǔguómǔqīn (祖国母亲 or 祖國母親, "ancestral land, the mother") is frequently used. the Czechs as vlast, power or (rarely) otčina, fatherland the Hungarians as szülőföld (literally: "bearing land" or "parental land") the Indians as मातृभूमि literally meaning "motherland" the Kurds as warê bav û kalan meaning "land of the fathers and the grandfathers" the Japanese as sokoku (祖国, "land of ancestors") the Koreans as joguk (조국, Hanja: 祖國, "land of ancestors") French speakers: Patrie, although they also use la mère patrie, which includes the idea of motherland the Latvians as tēvija or tēvzeme (although dzimtene – roughly translated as "place that somebody grew up" – is more neutral and used more commonly nowadays) the Burmese as အမိမြေ (ami-myay) literally meaning "motherland" the Persians as Sarzamin e Pedari (Fatherland), Sarzamin e Mādari (Motherland) or Mihan (Home) the Poles as ojczyzna (ojczyzna is derived from ojciec, Polish for father, but ojczyzna itself and Polska are feminine, so it can also be translated as motherland), also an archaism macierz "mother" is rarely used the Russians, as Otechestvo (отечество) or Otchizna (отчизна), both words derived from отец, Russian for father. Otechestvo is neuter, otchizna is feminine. the Slovenes as očetnjava, although domovina (homeland) is more common. the Swedes as fäderneslandet, although fosterlandet is more common (meaning the land that fostered/raised a person) the Vietnamese as Tổ quốc (Chữ Nôm: 祖國, "land of ancestors") Uses by country The Soviet Union created homelands for some minorities in the 1920s, including the Volga German ASSR and the Jewish Autonomous Oblast. In the case of the Volga German ASSR, these homelands were later abolished and their inhabitants deported to either Siberia or the Kazakh SSR. In the United States, the Department of Homeland Security was created soon after the 11 September 2001, terrorist attacks, as a means to centralize response to various threats. In a June 2002 column, Republican consultant and speechwriter Peggy Noonan expressed the hope that the Bush administration would change the name of the department, writing that, "The name Homeland Security grates on a lot of people, understandably. Homeland isn't really an American word, it's not something we used to say or say now". In the apartheid era in South Africa, the concept was given a different meaning. The white government had designated approximately 25% of its non-desert territory for black tribal settlement. Whites and other non-blacks were restricted from owning land or settling in those areas. After 1948 they were gradually granted an increasing level of "home-rule". From 1976 several of these regions were granted independence. Four of them were declared independent nations by South Africa, but were unrecognized as independent countries by any other nation besides each other and South Africa. The territories set aside for the African inhabitants were also known as bantustans. In Australia, the term refers to relatively small Aboriginal settlements (referred to also as "outstations") where people with close kinship ties share lands significant to them for cultural reasons. Many such homelands are found across Western Australia, the Northern Territory, and Queensland. The homeland movement gained momentum in the 1970 and 1980s. Not all homelands are permanently occupied owing to seasonal or cultural reasons. Much of their funding and support have been withdrawn since the 2000s. In Turkish, the concept of "homeland", especially in the patriotic sense, is "ana vatan" (lit. mother homeland), while "baba ocağı" (lit. father's hearth) is used to refer to one's childhood home. (Note: The Turkish word "ocak" has the double meaning of january and fireplace, like the Spanish "hogar", which can mean "home" or "hearth".) Land of one's home In some languages, there are additional words that refer specifically to the place where one is home to, but is narrower in scope than one's nation, and often have some sort of nostalgic, fantastic, heritage connection, for example: In German language, . In Japanese language, , or . In Chinese languages, or . In Vietnamese language, . In Korean language, , . See also Diaspora politics Heimat Homeland security Mother tongue Separatism Secession References Further reading Landscape and Memory by Simon Schama (Random House, 1995) External links Nationalism and Ethnicity – A Theoretical Overview Nationalism Cultural geography Ethnicity in politics
14082
https://en.wikipedia.org/wiki/Horse%20breeding
Horse breeding
Horse breeding is reproduction in horses, and particularly the human-directed process of selective breeding of animals, particularly purebred horses of a given breed. Planned matings can be used to produce specifically desired characteristics in domesticated horses. Furthermore, modern breeding management and technologies can increase the rate of conception, a healthy pregnancy, and successful foaling. Terminology The male parent of a horse, a stallion, is commonly known as the sire and the female parent, the mare, is called the dam. Both are genetically important, as each parent genes can be existent with a 50% probability in the foal. Contrary to popular misuse, "colt" refers to a young male horse only; "filly" is a young female. Though many horse owners may simply breed a family mare to a local stallion in order to produce a companion animal, most professional breeders use selective breeding to produce individuals of a given phenotype, or breed. Alternatively, a breeder could, using individuals of differing phenotypes, create a new breed with specific characteristics. A horse is "bred" where it is foaled (born). Thus a colt conceived in England but foaled in the United States is regarded as being bred in the US. In some cases, most notably in the Thoroughbred breeding industry, American- and Canadian-bred horses may also be described by the state or province in which they are foaled. Some breeds denote the country, or state, where conception took place as the origin of the foal. Similarly, the "breeder", is the person who owned or leased the mare at the time of foaling. That individual may not have had anything to do with the mating of the mare. It is important to review each breed registry's rules to determine which applies to any specific foal. In the horse breeding industry, the term "half-brother" or "half-sister" only describes horses which have the same dam, but different sires. Horses with the same sire but different dams are simply said to be "by the same sire", and no sibling relationship is implied. "Full" (or "own") siblings have both the same dam and the same sire. The terms paternal half-sibling, and maternal half-sibling are also often used. Three-quarter siblings are horses out of the same dam, and are by sires that are either half-brothers (i.e. same dam) or who are by the same sire. Thoroughbreds and Arabians are also classified through the "distaff" or direct female line, known as their "family" or "tail female" line, tracing back to their taproot foundation bloodstock or the beginning of their respective stud books. The female line of descent always appears at the bottom of a tabulated pedigree and is therefore often known as the bottom line. In addition, the maternal grandfather of a horse has a special term: damsire. "Linebreeding" technically is the duplication of fourth-generation or more distant ancestors. However, the term is often used more loosely, describing horses with duplication of ancestors closer than the fourth generation. It also is sometimes used as a euphemism for the practice of inbreeding, a practice that is generally frowned upon by horse breeders, though used by some in an attempt to fix certain traits. Estrous cycle of the mare The estrous cycle (also spelled oestrous) controls when a mare is sexually receptive toward a stallion, and helps to physically prepare the mare for conception. It generally occurs during the spring and summer months, although some mares may be sexually receptive into the late fall, and is controlled by the photoperiod (length of the day), the cycle first triggered when the days begin to lengthen. The estrous cycle lasts about 19–22 days, with the average being 21 days. As the days shorten, the mare returns to a period when she is not sexually receptive, known as anestrus. Anestrus – occurring in the majority of, but not all, mares – prevents the mare from conceiving in the winter months, as that would result in her foaling during the harshest part of the year, a time when it would be most difficult for the foal to survive. This cycle contains 2 phases: Estrus, or Follicular, phase: 5–7 days in length, when the mare is sexually receptive to a stallion. Estrogen is secreted by the follicle. Ovulation occurs in the final 24–48 hours of estrus. Diestrus, or Luteal, phase: 14–15 days in length, the mare is not sexually receptive to the stallion. The corpus luteum secretes progesterone. Depending on breed, on average, 16% of mares have double ovulations, allowing them to twin, though this does not affect the length of time of estrus or diestrus. Effects on the reproductive system during the estrous cycle Changes in hormone levels can have great effects on the physical characteristics of the reproductive organs of the mare, thereby preparing, or preventing, her from conceiving. Uterus: increased levels of estrogen during estrus cause edema within the uterus, making it feel heavier, and the uterus loses its tone. This edema decreases following ovulation, and the muscular tone increases. High levels of progesterone do not cause edema within the uterus. The uterus becomes flaccid during anestrus. Cervix: the cervix starts to relax right before estrus occurs, with maximal relaxation around the time of ovulation. The secretions of the cervix increase. High progesterone levels (during diestrus) cause the cervix to close and become toned. Vagina: the portion of the vagina near the cervix becomes engorged with blood right before estrus. The vagina becomes relaxed and secretions increase. Vulva: relaxes right before estrus begins. Becomes dry, and closes more tightly, during diestrus. Hormones involved in the estrous cycle, during foaling, and after birth The cycle is controlled by several hormones which regulate the estrous cycle, the mare's behavior, and the reproductive system of the mare. The cycle begins when the increased day length causes the pineal gland to reduce the levels of melatonin, thereby allowing the hypothalamus to secrete GnRH. GnRH (Gonadotropin releasing hormone): secreted by the hypothalamus, causes the pituitary to release two gonadotrophins: LH and FSH. LH (Luteinizing hormone): levels are highest 2 days following ovulation, then slowly decrease over 4–5 days, dipping to their lowest levels 5–16 days after ovulation. Stimulates maturation of the follicle, which then in turn secretes estrogen. Unlike most mammals, the mare does not have an increase of LH right before ovulation. FSH (Follicle-stimulating hormone): secreted by the pituitary, causes the ovarian follicle to develop. Levels of FSH rise slightly at the end of estrus, but have their highest peak about 10 days before the next ovulation. FSH is inhibited by inhibin (see below), at the same time LH and estrogen levels rise, which prevents immature follicles from continuing their growth. Mares may however have multiple FSH waves during a single estrous cycle, and diestrus follicles resulting from a diestrus FSH wave are not uncommon, particularly in the height of the natural breeding season. Estrogen: secreted by the developing follicle, it causes the pituitary gland to secrete more LH (therefore, these 2 hormones are in a positive feedback loop). Additionally, it causes behavioral changes in the mare, making her more receptive toward the stallion, and causes physical changes in the cervix, uterus, and vagina to prepare the mare for conception (see above). Estrogen peaks 1–2 days before ovulation, and decreases within 2 days following ovulation. Inhibin: secreted by the developed follicle right before ovulation, "turns off" FSH, which is no longer needed now that the follicle is larger. Progesterone: prevents conception and decreases sexual receptibility of the mare to the stallion. Progesterone is therefore lowest during the estrus phase, and increases during diestrus. It decreases 12–15 days after ovulation, when the corpus luteum begins to decrease in size. Prostaglandin: secreted by the endrometrium 13–15 days following ovulation, causes luteolysis and prevents the corpus luteum from secreting progesterone eCG – equine chorionic gonadotropin – also called PMSG (pregnant mare serum gonadotropin): chorionic gonadotropins secreted if the mare conceives. First secreted by the endometrial cups around the 36th day of gestation, peaking around day 60, and decreasing after about 120 days of gestation. Also help to stimulate the growth of the fetal gonads. Prolactin: stimulates lactation Oxytocin: stimulates the uterus to contract Breeding and gestation While horses in the wild mate and foal in mid to late spring, in the case of horses domestically bred for competitive purposes, especially horse racing, it is desirable that they be born as close to January 1 in the northern hemisphere or August 1 in the southern hemisphere as possible, so as to be at an advantage in size and maturity when competing against other horses in the same age group. When an early foal is desired, barn managers will put the mare "under lights" by keeping the barn lights on in the winter to simulate a longer day, thus bringing the mare into estrus sooner than she would in nature. Mares signal estrus and ovulation by urination in the presence of a stallion, raising the tail and revealing the vulva. A stallion, approaching with a high head, will usually nicker, nip and nudge the mare, as well as sniff her urine to determine her readiness for mating. Once fertilized, the oocyte (egg) remains in the oviduct for approximately 5.5 more days, and then descends into the uterus. The initial single cell combination is already dividing and by the time of entry into the uterus, the egg might have already reached the blastocyst stage. The gestation period lasts for about eleven months, or about 340 days (normal average range 320–370 days). During the early days of pregnancy, the conceptus is mobile, moving about in the uterus until about day 16 when "fixation" occurs. Shortly after fixation, the embryo proper (so called up to about 35 days) will become visible on trans-rectal ultrasound (about day 21) and a heartbeat should be visible by about day 23. After the formation of the endometrial cups and early placentation is initiated (35–40 days of gestation) the terminology changes, and the embryo is referred to as a fetus. True implantation – invasion into the endometrium of any sort – does not occur until about day 35 of pregnancy with the formation of the endometrial cups, and true placentation (formation of the placenta) is not initiated until about day 40-45 and not completed until about 140 days of pregnancy. The fetus's sex can be determined by day 70 of the gestation using ultrasound. Halfway through gestation the fetus is the size of between a rabbit and a beagle. The most dramatic fetal development occurs in the last 3 months of pregnancy when 60% of fetal growth occurs. Colts are carried on average about 4 days longer than fillies. Care of the pregnant mare Domestic mares receive specific care and nutrition to ensure that they and their foals are healthy. Mares are given vaccinations against diseases such as the Rhinopneumonitis (EHV-1) virus (which can cause miscarriage) as well as vaccines for other conditions that may occur in a given region of the world. Pre-foaling vaccines are recommended 4–6 weeks prior to foaling to maximize the immunoglobulin content of the colostrum in the first milk. Mares are dewormed a few weeks prior to foaling, as the mare is the primary source of parasites for the foal. Mares can be used for riding or driving during most of their pregnancy. Exercise is healthy, though should be moderated when a mare is heavily in foal. Exercise in excessively high temperatures has been suggested as being detrimental to pregnancy maintenance during the embryonic period; however ambient temperatures encountered during the research were in the region of 100 degrees F and the same results may not be encountered in regions with lower ambient temperatures. During the first several months of pregnancy, the nutritional requirements do not increase significantly since the rate of growth of the fetus is very slow. However, during this time, the mare may be provided supplemental vitamins and minerals, particularly if forage quality is questionable. During the last 3–4 months of gestation, rapid growth of the fetus increases the mare's nutritional requirements. Energy requirements during these last few months, and during the first few months of lactation are similar to those of a horse in full training. Trace minerals such as copper are extremely important, particularly during the tenth month of pregnancy, for proper skeletal formation. Many feeds designed for pregnant and lactating mares provide the careful balance required of increased protein, increased calories through extra fat as well as vitamins and minerals. Overfeeding the pregnant mare, particularly during early gestation, should be avoided, as excess weight may contribute to difficulties foaling or fetal/foal related problems. Foaling Mares due to foal are usually separated from other horses, both for the benefit of the mare and the safety of the soon-to-be-delivered foal. In addition, separation allows the mare to be monitored more closely by humans for any problems that may occur while giving birth. In the northern hemisphere, a special foaling stall that is large and clutter free is frequently used, particularly by major breeding farms. Originally, this was due in part to a need for protection from the harsh winter climate present when mares foal early in the year, but even in moderate climates, such as Florida, foaling stalls are still common because they allow closer monitoring of mares. Smaller breeders often use a small pen with a large shed for foaling, or they may remove a wall between two box stalls in a small barn to make a large stall. In the milder climates seen in much of the southern hemisphere, most mares foal outside, often in a paddock built specifically for foaling, especially on the larger stud farms. Many stud farms worldwide employ technology to alert human managers when the mare is about to foal, including webcams, closed-circuit television, or assorted types of devices that alert a handler via a remote alarm when a mare lies down in a position to foal. On the other hand, some breeders, particularly those in remote areas or with extremely large numbers of horses, may allow mares to foal out in a field amongst a herd, but may also see higher rates of foal and mare mortality in doing so. Most mares foal at night or early in the morning, and prefer to give birth alone when possible. Labor is rapid, often no more than 30 minutes, and from the time the feet of the foal appear to full delivery is often only about 15 to 20 minutes. Once the foal is born, the mare will lick the newborn foal to clean it and help blood circulation. In a very short time, the foal will attempt to stand and get milk from its mother. A foal should stand and nurse within the first hour of life. To create a bond with her foal, the mare licks and nuzzles the foal, enabling her to distinguish the foal from others. Some mares are aggressive when protecting their foals, and may attack other horses or unfamiliar humans that come near their newborns. After birth, a foal's navel is dipped in antiseptic to prevent infection. The foal is sometimes given an enema to help clear the meconium from its digestive tract. The newborn is monitored to ensure that it stands and nurses without difficulty. While most horse births happen without complications, many owners have first aid supplies prepared and a veterinarian on call in case of a birthing emergency. People who supervise foaling should also watch the mare to be sure that she passes the placenta in a timely fashion, and that it is complete with no fragments remaining in the uterus. Retained fetal membranes can cause a serious inflammatory condition (endometritis) and/or infection. If the placenta is not removed from the stall after it is passed, a mare will often eat it, an instinct from the wild, where blood would attract predators. Foal care Foals develop rapidly, and within a few hours a wild foal can travel with the herd. In domestic breeding, the foal and dam are usually separated from the herd for a while, but within a few weeks are typically pastured with the other horses. A foal will begin to eat hay, grass and grain alongside the mare at about 4 weeks old; by 10–12 weeks the foal requires more nutrition than the mare's milk can supply. Foals are typically weaned at 4–8 months of age, although in the wild a foal may nurse for a year. How breeds develop Beyond the appearance and conformation of a specific type of horse, breeders aspire to improve physical performance abilities. This concept, known as matching "form to function," has led to the development of not only different breeds, but also families or bloodlines within breeds that are specialists for excelling at specific tasks. For example, the Arabian horse of the desert naturally developed speed and endurance to travel long distances and survive in a harsh environment, and domestication by humans added a trainable disposition to the animal's natural abilities. In the meantime, in northern Europe, the locally adapted heavy horse with a thick, warm coat was domesticated and put to work as a farm animal that could pull a plow or wagon. This animal was later adapted through selective breeding to create a strong but rideable animal suitable for the heavily armored knight in warfare. Then, centuries later, when people in Europe wanted faster horses than could be produced from local horses through simple selective breeding, they imported Arabians and other oriental horses to breed as an outcross to the heavier, local animals. This led to the development of breeds such as the Thoroughbred, a horse taller than the Arabian and faster over the distances of a few miles required of a European race horse or light cavalry horse. Another cross between oriental and European horses produced the Andalusian, a horse developed in Spain that was powerfully built, but extremely nimble and capable of the quick bursts of speed over short distances necessary for certain types of combat as well as for tasks such as bullfighting. Later, the people who settled America needed a hardy horse that was capable of working with cattle. Thus, Arabians and Thoroughbreds were crossed on Spanish horses, both domesticated animals descended from those brought over by the Conquistadors, and feral horses such as the Mustangs, descended from the Spanish horse, but adapted by natural selection to the ecology and climate of the west. These crosses ultimately produced new breeds such as the American Quarter Horse and the Criollo of Argentina. In Canada, the Canadian Horse descended from the French stock Louis XIV sent to Canada in the late 17th century.[6] The initial shipment, in 1665, consisted of two stallions and twenty mares from the Royal Stables in Normandy and Brittany, the centre of French horse breeding.[7] Only 12 of the 20 mares survived the trip. Two more shipments followed, one in 1667 of 14 horses (mostly mares, but with at least one stallion), and one in 1670 of 11 mares and a stallion. The shipments included a mix of draft horses and light horses, the latter of which included both pacing and trotting horses.[1] The exact origins of all the horses are unknown, although the shipments probably included Bretons, Normans, Arabians, Andalusians and Barbs. In modern times, these breeds themselves have since been selectively bred to further specialize at certain tasks. One example of this is the American Quarter Horse. Once a general-purpose working ranch horse, different bloodlines now specialize in different events. For example, larger, heavier animals with a very steady attitude are bred to give competitors an advantage in events such as team roping, where a horse has to start and stop quickly, but also must calmly hold a full-grown steer at the end of a rope. On the other hand, for an event known as cutting, where the horse must separate a cow from a herd and prevent it from rejoining the group, the best horses are smaller, quick, alert, athletic and highly trainable. They must learn quickly, have conformation that allows quick stops and fast, low turns, and the best competitors have a certain amount of independent mental ability to anticipate and counter the movement of a cow, popularly known as "cow sense." Another example is the Thoroughbred. While most representatives of this breed are bred for horse racing, there are also specialized bloodlines suitable as show hunters or show jumpers. The hunter must have a tall, smooth build that allows it to trot and canter smoothly and efficiently. Instead of speed, value is placed on appearance and upon giving the equestrian a comfortable ride, with natural jumping ability that shows bascule and good form. A show jumper, however, is bred less for overall form and more for power over tall fences, along with speed, scope, and agility. This favors a horse with a good galloping stride, powerful hindquarters that can change speed or direction easily, plus a good shoulder angle and length of neck. A jumper has a more powerful build than either the hunter or the racehorse. History of horse breeding The history of horse breeding goes back millennia. Though the precise date is in dispute, humans could have domesticated the horse as far back as approximately 4500 BCE. However, evidence of planned breeding has a more blurry history. It is well known, for example, that the Romans did breed horses and valued them in their armies, but little is known regarding their breeding and husbandry practices: all that remains are statues and artwork. Mankind has plenty of equestrian statues of Roman emperors, horses are mentioned in the Odyssey by Homer, and hieroglyphics and paintings left behind by Egyptians tell stories of pharaohs hunting elephants from chariots. Nearly nothing is known of what became of the horses they bred for hippodromes, for warfare, or even for farming. One of the earliest people known to document the breedings of their horses were the Bedouin of the Middle East, the breeders of the Arabian horse. While it is difficult to determine how far back the Bedouin passed on pedigree information via an oral tradition, there were written pedigrees of Arabian horses by CE 1330. The Akhal-Teke of West-Central Asia is another breed with roots in ancient times that was also bred specifically for war and racing. The nomads of the Mongolian steppes bred horses for several thousand years as well, and the Caspian horse is believed to be a very close relative of Ottoman horses from the earliest origins of the Turks in Central Asia. The types of horse bred varied with culture and with the times. The uses to which a horse was put also determined its qualities, including smooth amblers for riding, fast horses for carrying messengers, heavy horses for plowing and pulling heavy wagons, ponies for hauling cars of ore from mines, packhorses, carriage horses and many others. Medieval Europe bred large horses specifically for war, called destriers. These horses were the ancestors of the great heavy horses of today, and their size was preferred not simply because of the weight of the armor, but also because a large horse provided more power for the knight's lance. Weighing almost twice as much as a normal riding horse, the destrier was a powerful weapon in battle meant to act like a giant battering ram that could quite literally run down men on an enemy line. On the other hand, during this same time, lighter horses were bred in northern Africa and the Middle East, where a faster, more agile horse was preferred. The lighter horse suited the raids and battles of desert people, allowing them to outmaneuver rather than overpower the enemy. When Middle Eastern warriors and European knights collided in warfare, the heavy knights were frequently outmaneuvered. The Europeans, however, responded by crossing their native breeds with "oriental" type horses such as the Arabian, Barb, and Turkoman horse This cross-breeding led both to a nimbler war horse, such as today's Andalusian horse, but also created a type of horse known as a Courser, a predecessor to the Thoroughbred, which was used as a message horse. During the Renaissance, horses were bred not only for war, but for haute ecole riding, derived from the most athletic movements required of a war horse, and popular among the elite nobility of the time. Breeds such as the Lipizzan and the now extinct Neapolitan horse were developed from Spanish-bred horses for this purpose, and also became the preferred mounts of cavalry officers, who were derived mostly from the ranks of the nobility. It was during this time that firearms were developed, and so the light cavalry horse, a faster and quicker war horse, was bred for "shoot and run" tactics rather than the shock action as in the Middle Ages. Fine horses usually had a well muscled, curved neck, slender body, and sweeping mane, as the nobility liked to show off their wealth and breeding in paintings of the era. After Charles II retook the British throne in 1660, horse racing, which had been banned by Cromwell, was revived. The Thoroughbred was developed 40 years later, bred to be the ultimate racehorse, through the lines of three foundation Arabian stallions and one Turkish horse. In the 18th century, James Burnett, Lord Monboddo noted the importance of selecting appropriate parentage to achieve desired outcomes of successive generations. Monboddo worked more broadly in the abstract thought of species relationships and evolution of species. The Thoroughbred breeding hub in Lexington, Kentucky was developed in the late 18th century, and became a mainstay in American racehorse breeding. The 17th and 18th centuries saw more of a need for fine carriage horses in Europe, bringing in the dawn of the warmblood. The warmblood breeds have been exceptionally good at adapting to changing times, and from their carriage horse beginnings they easily transitioned during the 20th century into a sport horse type. Today's warmblood breeds, although still used for competitive driving, are more often seen competing in show jumping or dressage. The Thoroughbred continues to dominate the horse racing world, although its lines have been more recently used to improve warmblood breeds and to develop sport horses. The French saddle horse is an excellent example as is the Irish Sport Horse, the latter being an unusual combination between a Thoroughbred and a draft breed. The American Quarter Horse was developed early in the 18th century, mainly for quarter racing (racing ¼ of a mile). Colonists did not have racetracks or any of the trappings of Europe that the earliest Thoroughbreds had at their disposal, so instead the owners of Quarter Horses would run their horses on roads that lead through town as a form of local entertainment. As the USA expanded West, the breed went with settlers as a farm and ranch animal, and "cow sense" was particularly valued: their use for herding cattle increased on rough, dry terrain that often involved sitting in the saddle for long hours. However, this did not mean that the original ¼-mile races that colonists held ever went out of fashion, so today there are three types: the stock horse type, the racer, and the more recently evolving sport type. The racing type most resembles the finer-boned ancestors of the first racing Quarter Horses, and the type is still used for ¼-mile races. The stock horse type, used in western events and as a farm and patrol animal is bred for a shorter stride, an ability to stop and turn quickly, and an unflappable attitude that remains calm and focused even in the face of an angry charging steer. The first two are still to this day bred to have a combination of explosive speed that exceeds the Thoroughbred on short distances clocked as high as 55 mph, but they still retain the gentle, calm, and kindly temperament of their ancestors that makes them easily handled. The Canadian horse's origin corresponds to shipments of French horses, some of which came from Louis XIV's own stable and most likely were Baroque horses meant to be gentlemen's mounts. These were ill-suited to farm work and to the hardscrabble life of the New World, so like the Americans, early Canadians crossed their horses with natives escapees. In time they evolved along similar lines as the Quarter Horse to the South as both the US and Canada spread westward and needed a calm and tractable horse versatile enough to carry the farmer's son to school but still capable of running fast and running hard as a cavalry horse, a stockhorse, or a horse to pull a conestoga wagon. Other horses from North America retained a hint of their mustang origins by being either derived from stock that Native Americans bred that came in a rainbow of color, like the Appaloosa and American Paint Horse. with those East of the Mississippi River increasingly bred to impress and mimic the trends of the upper classes of Europe: The Tennessee Walking Horse and Saddlebred were originally plantation horses bred for their gait and comfortable ride in the saddle as a plantation master would survey his vast lands like an English lord. Horses were needed for heavy draft and carriage work until replaced by the automobile, truck, and tractor. After this time, draft and carriage horse numbers dropped significantly, though light riding horses remained popular for recreational pursuits. Draft horses today are used on a few small farms, but today are seen mainly for pulling and plowing competitions rather than farm work. Heavy harness horses are now used as an outcross with lighter breeds, such as the Thoroughbred, to produce the modern warmblood breeds popular in sport horse disciplines, particularly at the Olympic level. Deciding to breed a horse Breeding a horse is an endeavor where the owner, particularly of the mare, will usually need to invest considerable time and money. For this reason, a horse owner needs to consider several factors, including: Does the proposed breeding animal have valuable genetic qualities to pass on? Is the proposed breeding animal in good physical health, fertile, and able to withstand the rigors of reproduction? For what purpose will the foal be used? Is there a market for the foal if the owner does not wish to keep the foal for its entire life? What is the anticipated economic benefit, if any, to the owner of the ensuing foal? What is the anticipated economic benefit, if any, to the owner(s) of the sire and dam or the foal? Does the owner of the mare have the expertise to properly manage the mare through gestation and parturition? Does the owner of the potential foal have the expertise to properly manage and train a young animal once it is born? There are value judgements involved in considering whether an animal is suitable breeding stock, hotly debated by breeders. Additional personal beliefs may come into play when considering a suitable level of care for the mare and ensuing foal, the potential market or use for the foal, and other tangible and intangible benefits to the owner. If the breeding endeavor is intended to make a profit, there are additional market factors to consider, which may vary considerably from year to year, from breed to breed, and by region of the world. In many cases, the low end of the market is saturated with horses, and the law of supply and demand thus allows little or no profit to be made from breeding unregistered animals or animals of poor quality, even if registered. The minimum cost of breeding for a mare owner includes the stud fee, and the cost of proper nutrition, management and veterinary care of the mare throughout gestation, parturition, and care of both mare and foal up to the time of weaning. Veterinary expenses may be higher if specialized reproductive technologies are used or health complications occur. Making a profit in horse breeding is often difficult. While some owners of only a few horses may keep a foal for purely personal enjoyment, many individuals breed horses in hopes of making some money in the process. A rule of thumb is that a foal intended for sale should be worth three times the cost of the stud fee if it were sold at the moment of birth. From birth forward, the costs of care and training are added to the value of the foal, with a sale price going up accordingly. If the foal wins awards in some form of competition, that may also enhance the price. On the other hand, without careful thought, foals bred without a potential market for them may wind up being sold at a loss, and in a worst-case scenario, sold for "salvage" value—a euphemism for sale to slaughter as horsemeat. Therefore, a mare owner must consider their reasons for breeding, asking hard questions of themselves as to whether their motivations are based on either emotion or profit and how realistic those motivations may be. Choosing breeding stock The stallion should be chosen to complement the mare, with the goal of producing a foal that has the best qualities of both animals, yet avoids having the weaker qualities of either parent. Generally, the stallion should have proven himself in the discipline or sport the mare owner wishes for the "career" of the ensuing foal. Mares should also have a competition record showing that they also have suitable traits, though this does not happen as often. Some breeders consider the quality of the sire to be more important than the quality of the dam. However, other breeders maintain that the mare is the most important parent. Because stallions can produce far more offspring than mares, a single stallion can have a greater overall impact on a breed. Research from Nagoya University supports the belief that the most important factor affecting a thoroughbred’s race performance is the quality of its sire, whereas the effect of the age of its broodmare is negligible. However, the mare may have a greater influence on an individual foal because its physical characteristics influence the developing foal in the womb and the foal also learns habits from its dam when young. Foals may also learn the "language of intimidation and submission" from their dam, and this imprinting may affect the foal's status and rank within the herd. Many times, a mature horse will achieve status in a herd similar to that of its dam; the offspring of dominant mares become dominant themselves. A purebred horse is usually worth more than a horse of mixed breeding, though this matters more in some disciplines than others. The breed of the horse is sometimes secondary when breeding for a sport horse, but some disciplines may prefer a certain breed or a specific phenotype of horse. Sometimes, purebred bloodlines are an absolute requirement: For example, most racehorses in the world must be recorded with a breed registry in order to race. Bloodlines are often considered, as some bloodlines are known to cross well with others. If the parents have not yet proven themselves by competition or by producing quality offspring, the bloodlines of the horse are often a good indicator of quality and possible strengths and weaknesses. Some bloodlines are known not only for their athletic ability, but could also carry a conformational or genetic defect, poor temperament, or for a medical problem. Some bloodlines are also fashionable or otherwise marketable, which is an important consideration should the mare owner wish to sell the foal. Horse breeders also consider conformation, size and temperament. All of these traits are heritable, and will determine if the foal will be a success in its chosen discipline. The offspring, or "get", of a stallion are often excellent indicators of his ability to pass on his characteristics, and the particular traits he actually passes on. Some stallions are fantastic performers but never produce offspring of comparable quality. Others sire fillies of great abilities but not colts. At times, a horse of mediocre ability sires foals of outstanding quality. Mare owners also look into the question of if the stallion is fertile and has successfully "settled" (i.e. impregnated) mares. A stallion may not be able to breed naturally, or old age may decrease his performance. Mare care boarding fees and semen collection fees can be a major cost. Costs related to breeding Breeding a horse can be an expensive endeavor, whether breeding a backyard competition horse or the next Olympic medalist. Costs may include: The stud and booking fee Fees for collecting, handling, and transporting semen (if AI is used and semen is shipped) Mare exams: to determine if she is healthy enough to breed, to determine when she ovulates, and (if AI is used) to inseminate her Mare transport, care, and board if the mare is bred live cover at the stallion's residence Veterinary bills to keep the pregnant mare healthy while in foal Possible veterinary bills during pregnancy or foaling should something go wrong Veterinary bills for the foal for its first exam a few days following foaling Stud fees are determined by the quality of the stallion, his performance record, the performance record of his get (offspring), as well as the sport and general market that the animal is standing for. The highest stud fees are generally for racing Thoroughbreds, which may charge from two to three thousand dollars for a breeding to a new or unproven stallion, to several hundred thousand dollars for a breeding to a proven producer of stakes winners. Stallions in other disciplines often have stud fees that begin in the range of $1,000 to $3,000, with top contenders who produce champions in certain disciplines able to command as much as $20,000 for one breeding. The lowest stud fees to breed to a grade horse or an animal of low-quality pedigree may only be $100–$200, but there are trade-offs: the horse will probably be unproven, and likely to produce lower-quality offspring than a horse with a stud fee that is in the typical range for quality breeding stock. As a stallion's career, either performance or breeding, improves, his stud fee tends to increase in proportion. If one or two offspring are especially successful, winning several stakes races or an Olympic medal, the stud fee will generally greatly increase. Younger, unproven stallions will generally have a lower stud fee earlier on in their careers. To help decrease the risk of financial loss should the mare die or abort the foal while pregnant, many studs have a live foal guarantee (LFG) – also known as "no foal, free return" or "NFFR" - allowing the owner to have a free breeding to their stallion the next year. However, this is not offered for every breeding. Covering the mare There are two general ways to "cover" or breed the mare: Live cover: the mare is brought to the stallion's residence and is covered "live" in the breeding shed. She may also be turned out in a pasture with the stallion for several days to breed naturally ('pasture bred'). The former situation is often preferred, as it provides a more controlled environment, allowing the breeder to ensure that the mare was covered, and places the handlers in a position to remove the horses from one another should one attempt to kick or bite the other. Artificial Insemination (AI): the mare is inseminated by a veterinarian or an equine reproduction manager, using either fresh, cooled or frozen semen. After the mare is bred or artificially inseminated, she is checked using ultrasound 14–16 days later to see if she "took", and is pregnant. A second check is usually performed at 28 days. If the mare is not pregnant, she may be bred again during her next cycle. It is considered safe to breed a mare to a stallion of much larger size. Because of the mare's type of placenta and its attachment and blood supply, the foal will be limited in its growth within the uterus to the size of the mare's uterus, but will grow to its genetic potential after it is born. Test breedings have been done with draft horse stallions bred to small mares with no increase in the number of difficult births. Live cover When breeding live cover, the mare is usually boarded at the stud. She may be "teased" several times with a stallion that will not breed to her, usually with the stallion being presented to the mare over a barrier. Her reaction to the teaser, whether hostile or passive, is noted. A mare that is in heat will generally tolerate a teaser (although this is not always the case), and may present herself to him, holding her tail to the side. A veterinarian may also determine if the mare is ready to be bred, by ultrasound or palpating daily to determine if ovulation has occurred. Live cover can also be done in liberty on a paddock or on pasture, although due to safety and efficacy concerns, it is not common at professional breeding farms. When it has been determined that the mare is ready, both the mare and intended stud will be cleaned. The mare will then be presented to the stallion, usually with one handler controlling the mare and one or more handlers in charge of the stallion. Multiple handlers are preferred, as the mare and stallion can be easily separated should there be any trouble. The Jockey Club, the organization that oversees the Thoroughbred industry in the United States, requires all registered foals to be bred through live cover. Artificial insemination, listed below, is not permitted. Similar rules apply in other countries, such as Australia. By contrast, the U.S. standardbred industry allows registered foals to be bred by live cover, or by artificial insemination (AI) with fresh or frozen (not dried) semen. No other artificial fertility treatment is allowed. In addition, foals bred via AI of frozen semen may only be registered if the stallion's sperm was collected during his lifetime, and used no later than the calendar year of his death or castration. Artificial insemination Whereas the various national Thoroughbred associations typically require live cover, by 2009 most horse breeds allowed for the artificial insemination of mares with cooled, frozen or even fresh semen. Artificial insemination (AI) has several advantages over live cover, and has a very similar conception rate: The mare and stallion never have to come in contact with each other, which therefore reduces breeding accidents, such as the mare kicking the stallion. AI opens up the world to international breeding, as semen may be shipped across continents to mares that would otherwise be unable to breed to a particular stallion. A mare also does not have to travel to the stallion, so the process is less stressful on her, and if she already has a foal, the foal does not have to travel. AI allows more mares to be bred from one stallion, as the ejaculate may be split between mares. AI reduces the chance of spreading sexually transmitted diseases between mare and stallion. AI allows mares or stallions with health issues, such as sore hocks which may prevent a stallion from mounting, to continue to breed. Frozen semen may be stored and used to breed mares even after the stallion is dead, allowing his lines to continue. However, the semen of some stallions does not freeze well. Some breed registries may not permit the registration of foals resulting from the use of frozen semen after the stallion's death, although other large registries accept such usage and provide registrations. The overall trend is toward permitting use of frozen semen after the death of the stallion. A stallion is usually trained to mount a phantom (or dummy) mare, although a live mare may be used, and he is most commonly collected using an artificial vagina (AV) which is heated to simulate the vagina of the mare. The AV has a filter and collection area at one end to capture the semen, which can then be processed in a lab. The semen may be chilled or frozen and shipped to the mare owner or used to breed mares "on-farm". When the mare is in heat, the person inseminating introduces the semen directly into her uterus using a syringe and pipette. Advanced reproductive techniques Often an owner does not want to take a valuable competition mare out of training to carry a foal. This presents a problem, as the mare will usually be quite old by the time she is retired from her competitive career, at which time it is more difficult to impregnate her. Other times, a mare may have physical problems that prevent or discourage breeding. However, there are now several options for breeding these mares. These options also allow a mare to produce multiple foals each breeding season, instead of the usual one. Therefore, mares may have an even greater value for breeding. Embryo transfer: This relatively new method involves flushing out the mare's fertilized embryo a few days following insemination, and transferring to a surrogate mare, which has been synchronized to be in the same phase of the estrous cycle as the donor mare. Gamete intrafallopian transfer (GIFT): The mare's ovum and the stallion's sperm are deposited in the oviduct of a surrogate dam. This technique is very useful for subfertile stallions, as fewer sperm are needed, so a stallion with a low sperm count can still successfully breed. Egg transfer: An oocyte is removed from the mare's follicle and transferred into the oviduct of the recipient mare, who is then bred. This is best for mares with physical problems, such as an obstructed oviduct, that prevent breeding. Intracytoplasmic sperm injection (ICSI): Used in horses due to lack of successful co-incubation of female and male gametes in simple IVF. A plug of the zona pellucida is removed and a single sperm cell is injected into the ooplasm of the mature oocyte. An advantage of ICSI over IVF is that lower quality sperm can be used since the sperm does not have to penetrate the zona pellucida. The success rate of ICSI is 23-44% blastocyst development. The world's first cloned horse, Prometea, was born in 2003. Other notable instances of horse cloning are: In 2006, Scamper, an extremely successful barrel racing horse, a gelding, was cloned. The resulting stallion, Clayton, became the first cloned horse to stand at stud in the U.S. In 2007, a renowned show jumper and Thoroughbred, Gem Twist, was cloned by Frank Chapot and his family. In September 2008, Gemini was born and several other clones followed, leading to the development of a breeding line from Gem Twist. In 2010, the first lived equine cloned of a Criollo horse was born in Argentina, and was the first horse clone produced in Latin America. In the same year a cloned polo horse was sold for $800,000 - the highest known price ever paid for a polo horse. In 2013, the world-famous polo star Adolfo Cambiaso helped his high-handicap team La Dolfina win the Argentine National Open, scoring nine goals in the 16-11 match. Two of those he scored atop a horse named Show Me, a clone, and the first to ride onto the Argentine pitch. See also Domestication of the horse Endometrosis Evolution of the horse Glossary of equestrian terms Pedigree chart Thoroughbred breeding theories References Further reading Riegal, Ronald J. DMV, and Susan E. Hakola DMV. Illustrated Atlas of Clinical Equine Anatomy and Common Disorders of the Horse Vol. II. Equistar Publication, Limited. Marysville, OH. Copyright 2000. Horse health Horse-related professions and professionals
14257
https://en.wikipedia.org/wiki/Haddocks%27%20Eyes
Haddocks' Eyes
"Haddocks' Eyes" is the nickname of the name of a song sung by The White Knight from Lewis Carroll's 1871 novel Through the Looking-Glass, chapter VIII. "Haddocks' Eyes" is an example used to elaborate on the symbolic status of the concept of "name": a name as identification marker may be assigned to anything, including another name, thus introducing different levels of symbolization. It has been discussed in several works on logic and philosophy. Naming The White Knight explains to Alice a confusing nomenclature for the song. To summarize: The song's name is called Haddocks' Eyes The song's name is The Aged Aged Man The song is called Ways and Means The song is A-sitting on a Gate The complicated terminology distinguishing between 'the song, what the song is called, the name of the song, and what the name of the song is called' both uses and mentions the use–mention distinction. The song The White Knight sings the song to a tune he claims as his own invention, but which Alice recognises as "I give thee all, I can no more". By the time Alice heard it, she was already tired of poetry. The song parodies the plot, but not the style or metre, of "Resolution and Independence" by William Wordsworth. Upon the Lonely Moor Like "Jabberwocky," another poem published in Through the Looking Glass, "Haddocks’ Eyes" appears to have been revised over the course of many years. In 1856, Carroll published the following poem anonymously under the name Upon the Lonely Moor. It bears an obvious resemblance to "Haddocks' Eyes." See also Nonsense verse References British poems 1871 poems Alice's Adventures in Wonderland Nonsense poetry Parodies of literature
14400
https://en.wikipedia.org/wiki/History%20of%20science
History of science
The history of science covers the development of science from ancient times to the present. It encompasses all three major branches of science: natural, social, and formal. Science's earliest roots can be traced to Ancient Egypt and Mesopotamia around 3000 to 1200 BCE. These civilizations' contributions to mathematics, astronomy, and medicine influenced later Greek natural philosophy of classical antiquity, wherein formal attempts were made to provide explanations of events in the physical world based on natural causes. After the fall of the Western Roman Empire, knowledge of Greek conceptions of the world deteriorated in Latin-speaking Western Europe during the early centuries (400 to 1000 CE) of the Middle Ages, but continued to thrive in the Greek-speaking Eastern Roman (or Byzantine) Empire. Aided by translations of Greek texts, the Hellenistic worldview was preserved and absorbed into the Arabic-speaking Muslim world during the Islamic Golden Age. The recovery and assimilation of Greek works and Islamic inquiries into Western Europe from the 10th to 13th century revived the learning of natural philosophy in the West. Natural philosophy was transformed during the Scientific Revolution in 16th- to 17th-century Europe, as new ideas and discoveries departed from previous Greek conceptions and traditions. The New Science that emerged was more mechanistic in its worldview, more integrated with mathematics, and more reliable and open as its knowledge was based on a newly defined scientific method. More "revolutions" in subsequent centuries soon followed. The chemical revolution of the 18th century, for instance, introduced new quantitative methods and measurements for chemistry. In the 19th century, new perspectives regarding the conservation of energy, age of Earth, and evolution came into focus. And in the 20th century, new discoveries in genetics and physics laid the foundations for new subdisciplines such as molecular biology and particle physics. Moreover, industrial and military concerns as well as the increasing complexity of new research endeavors ushered in the era of "big science," particularly after the Second World War. Approaches to history of science The nature of the history of science (and by implication, the definition of science itself) is a topic of debate. The history of science is often seen as a linear story of progress but historians have shown that the story is more complex. Science is a human activity, and scientific contributions have been made by people from a wide range of different backgrounds and cultures. Science is increasingly seen as part of a global history of exchange, conflict and collaboration. The relationship between science and religion has been characterized in terms of "conflict", "harmony", "complexity", and "mutual independence", among others. Events in Europe such as the Galileo affair of the early 17th century, associated with the scientific revolution and the Age of Enlightenment, led scholars such as John William Draper to postulate () a conflict thesis, suggesting that religion and science have been in conflict methodologically, factually and politically throughout history. The conflict thesis has since lost favor among a majority contemporary scientists and historians of science. However, some contemporary philosophers and scientists, such as Richard Dawkins, Lawrence Krauss, Peter Atkins, and Donald Prothero, still subscribe to this thesis. Historians have emphasized that trust is necessary for claims about nature to be agreed upon. In this light, the establishment of the Royal Society and its code of experiment - trustworthy because witnessed by its members - has become an important chapter in the historiography of science. Many people in modern history (typically women and persons of color) were excluded from elite scientific communities and characterized by science as inferior. Historians in the 1980s and 1990s described the structural barriers to participation and began to recover the contributions of overlooked individuals. Historians have also investigated the mundane practices of science such as fieldwork and specimen collection, correspondence, drawing, record-keeping, and the use of laboratory and field equipment. Prehistoric times In prehistoric times, knowledge and technique were passed from generation to generation in an oral tradition. For instance, the domestication of maize for agriculture has been dated to about 9,000 years ago in southern Mexico, before the development of writing systems. Similarly, archaeological evidence indicates the development of astronomical knowledge in preliterate societies. The oral tradition of preliterate societies had several features, the first of which was its fluidity. New information was constantly absorbed and adjusted to new circumstances or community needs. There were no archives or reports. This fluidity was closely related to the practical need to explain and justify a present state of affairs. Another feature was the tendency to describe the universe as just sky and earth, with a potential underworld. They were also prone to identify causes with beginnings, thereby providing a historical origin with an explanation. There was also a reliance on a "medicine man" or "wise woman" for healing, knowledge of divine or demonic causes of diseases, and in more extreme cases, for rituals such as exorcism, divination, songs, and incantations. Finally, there was an inclination to unquestioningly accept explanations that might be deemed implausible in more modern times while at the same time not being aware that such credulous behaviors could have posed problems. The development of writing enabled humans to store and communicate knowledge across generations with much greater accuracy. Its invention was a prerequisite for the development of philosophy and later science in ancient times. Moreover, the extent to which philosophy and science would flourish in ancient times depended on the efficiency of a writing system (e.g., use of alphabets). Earliest roots The earliest roots of science can be traced to Ancient Egypt and Mesopotamia in around 3000 to 1200 BCE. Ancient Egypt Number system and geometry Starting in around 3000 BCE, the ancient Egyptians developed a numbering system that was decimal in character and had orientated their knowledge of geometry to solving practical problems such as those of surveyors and builders. They even developed an official calendar that contained twelve months, thirty days each, and five days at the end of the year. Their development of geometry was a necessary outgrowth of surveying to preserve the layout and ownership of farmland, which was flooded annually by the Nile river. The 3-4-5 right triangle and other rules of geometry were used to build rectilinear structures, and the post and lintel architecture of Egypt. Disease and healing Egypt was also a center of alchemy research for much of the Mediterranean. Based on the medical papyri written in the 2500–1200 BCE, the ancient Egyptians believed that disease was mainly caused by the invasion of bodies by evil forces or spirits. Thus, in addition to using medicines, their healing therapies included prayer, incantation, and ritual. The Ebers Papyrus, written in around 1600 BCE, contains medical recipes for treating diseases related to the eyes, mouths, skins, internal organs, and extremities as well as abscesses, wounds, burns, ulcers, swollen glands, tumors, headaches, and even bad breath. The Edwin Smith papyrus, written at about the same time, contains a surgical manual for treating wounds, fractures, and dislocations. The Egyptians believed that the effectiveness of their medicines depended on the preparation and administration under appropriate rituals. Medical historians believe that ancient Egyptian pharmacology, for example, was largely ineffective. Both the Ebers and Edwin Smith papyri applied the following components to the treatment of disease: examination, diagnosis, treatment, and prognosis, which display strong parallels to the basic empirical method of science and, according to G.E.R. Lloyd, played a significant role in the development of this methodology. Calendar The ancient Egyptians even developed an official calendar that contained twelve months, thirty days each, and five days at the end of the year. Unlike the Babylonian calendar or the ones used in Greek city-states at the time, the official Egyptian calendar was much simpler as it was fixed and did not take lunar and solar cycles into consideration. Mesopotamia The ancient Mesopotamians had extensive knowledge about the chemical properties of clay, sand, metal ore, bitumen, stone, and other natural materials, and applied this knowledge to practical use in manufacturing pottery, faience, glass, soap, metals, lime plaster, and waterproofing. Metallurgy required knowledge about the properties of metals. Nonetheless, the Mesopotamians seem to have had little interest in gathering information about the natural world for the mere sake of gathering information and were far more interested in studying the manner in which the gods had ordered the universe. Biology of non-human organisms was generally only written about in the context of mainstream academic disciplines. Animal physiology was studied extensively for the purpose of divination; the anatomy of the liver, which was seen as an important organ in haruspicy, was studied in particularly intensive detail. Animal behavior was also studied for divinatory purposes. Most information about the training and domestication of animals was probably transmitted orally without being written down, but one text dealing with the training of horses has survived. Mesopotamian medicine The ancient Mesopotamians had no distinction between "rational science" and magic. When a person became ill, doctors prescribed magical formulas to be recited as well as medicinal treatments. The earliest medical prescriptions appear in Sumerian during the Third Dynasty of Ur ( 2112 BCE – 2004 BCE). The most extensive Babylonian medical text, however, is the Diagnostic Handbook written by the ummânū, or chief scholar, Esagil-kin-apli of Borsippa, during the reign of the Babylonian king Adad-apla-iddina (1069–1046 BCE). In East Semitic cultures, the main medicinal authority was a kind of exorcist-healer known as an āšipu. The profession was generally passed down from father to son and was held in extremely high regard. Of less frequent recourse was another kind of healer known as an asu, who corresponds more closely to a modern physician and treated physical symptoms using primarily folk remedies composed of various herbs, animal products, and minerals, as well as potions, enemas, and ointments or poultices. These physicians, who could be either male or female, also dressed wounds, set limbs, and performed simple surgeries. The ancient Mesopotamians also practiced prophylaxis and took measures to prevent the spread of disease. Astronomy and celestial divination In Babylonian astronomy, records of the motions of the stars, planets, and the moon are left on thousands of clay tablets created by scribes. Even today, astronomical periods identified by Mesopotamian proto-scientists are still widely used in Western calendars such as the solar year and the lunar month. Using these data they developed arithmetical methods to compute the changing length of daylight in the course of the year and to predict the appearances and disappearances of the Moon and planets and eclipses of the Sun and Moon. Only a few astronomers' names are known, such as that of Kidinnu, a Chaldean astronomer and mathematician. Kiddinu's value for the solar year is in use for today's calendars. Babylonian astronomy was "the first and highly successful attempt at giving a refined mathematical description of astronomical phenomena." According to the historian A. Aaboe, "all subsequent varieties of scientific astronomy, in the Hellenistic world, in India, in Islam, and in the West—if not indeed all subsequent endeavor in the exact sciences—depend upon Babylonian astronomy in decisive and fundamental ways." To the Babylonians and other Near Eastern cultures, messages from the gods or omens were concealed in all natural phenomena that could be deciphered and interpreted by those who are adept. Hence, it was believed that the gods could speak through all terrestrial objects (e.g., animal entrails, dreams, malformed births, or even the color of a dog urinating on a person) and celestial phenomena. Moreover, Babylonian astrology was inseparable from Babylonian astronomy. Mathematics The Mesopotamian cuneiform tablet Plimpton 322, dating to the eighteenth-century BCE, records a number of Pythagorean triplets (3,4,5) (5,12,13) ..., hinting that the ancient Mesopotamians might have been aware of the Pythagorean theorem over a millennium before Pythagoras. Separate developments Mathematical achievements from Mesopotamia had some influence on the development of mathematics in India, and there were confirmed transmissions of mathematical ideas between India and China, which were bidirectional. Nevertheless, the mathematical and scientific achievements in India and particularly in China occurred largely independently from those of Europe and the confirmed early influences that these two civilizations had on the development of science in Europe in the pre-modern era were indirect, with Mesopotamia and later the Islamic World acting as intermediaries. The arrival of modern science, which grew out of the Scientific Revolution, in India and China and the greater Asian region in general can be traced to the scientific activities of Jesuit missionaries who were interested in studying the region's flora and fauna during the 16th to 17th century. India Indian astronomy and mathematics The earliest traces of mathematical knowledge in the Indian subcontinent appear with the Indus Valley Civilisation (c. 4th millennium BCE ~ c. 3rd millennium BCE). The people of this civilization made bricks whose dimensions were in the proportion 4:2:1, which is favorable for the stability of a brick structure. They also tried to standardize measurement of length to a high degree of accuracy. They designed a ruler—the Mohenjo-daro ruler—whose unit of length (approximately 1.32 inches or 3.4 centimeters) was divided into ten equal parts. Bricks manufactured in ancient Mohenjo-daro often had dimensions that were integral multiples of this unit of length. Indian astronomer and mathematician Aryabhata (476–550), in his Aryabhatiya (499) introduced the sine function in trigonometry. In 628 CE, Brahmagupta suggested that gravity was a force of attraction. He also lucidly explained the use of zero as both a placeholder and a decimal digit, along with the Hindu–Arabic numeral system now used universally throughout the world. Arabic translations of the two astronomers' texts were soon available in the Islamic world, introducing what would become Arabic numerals to the Islamic world by the 9th century. During the 14th–16th centuries, the Kerala school of astronomy and mathematics made significant advances in astronomy and especially mathematics, including fields such as trigonometry and analysis. In particular, Madhava of Sangamagrama is considered the "founder of mathematical analysis". In the Tantrasangraha treatise, Nilakantha Somayaji's updated the Aryabhatan model for the interior planets, Mercury, and Venus and the equation that he specified for the center of these planets was more accurate than the ones in European or Islamic astronomy until the time of Johannes Kepler in the 17th century. The first textual mention of astronomical concepts comes from the Vedas, religious literature of India. According to Sarma (2008): "One finds in the Rigveda intelligent speculations about the genesis of the universe from nonexistence, the configuration of the universe, the spherical self-supporting earth, and the year of 360 days divided into 12 equal parts of 30 days each with a periodical intercalary month.". The first 12 chapters of the Siddhanta Shiromani, written by Bhāskara in the 12th century, cover topics such as: mean longitudes of the planets; true longitudes of the planets; the three problems of diurnal rotation; syzygies; lunar eclipses; solar eclipses; latitudes of the planets; risings and settings; the moon's crescent; conjunctions of the planets with each other; conjunctions of the planets with the fixed stars; and the patas of the sun and moon. The 13 chapters of the second part cover the nature of the sphere, as well as significant astronomical and trigonometric calculations based on it. Grammar Some of the earliest linguistic activities can be found in Iron Age India (1st millennium BCE) with the analysis of Sanskrit for the purpose of the correct recitation and interpretation of Vedic texts. The most notable grammarian of Sanskrit was (c. 520–460 BCE), whose grammar formulates close to 4,000 rules for Sanskrit. Inherent in his analytic approach are the concepts of the phoneme, the morpheme and the root. The Tolkāppiyam text, composed in the early centuries of the common era, is a comprehensive text on Tamil grammar, which includes sutras on orthography, phonology, etymology, morphology, semantics, prosody, sentence structure and the significance of context in language. Medicine Findings from Neolithic graveyards in what is now Pakistan show evidence of proto-dentistry among an early farming culture. The ancient text Suśrutasamhitā of Suśruta describes procedures on various forms of surgery, including rhinoplasty, the repair of torn ear lobes, perineal lithotomy, cataract surgery, and several other excisions and other surgical procedures. Politics and state An ancient Indian treatise on statecraft, economic policy and military strategy by Kautilya and , who are traditionally identified with (c. 350–283 BCE). In this treatise, the behaviors and relationships of the people, the King, the State, the Government Superintendents, Courtiers, Enemies, Invaders, and Corporations are analyzed and documented. Roger Boesche describes the Arthaśāstra as "a book of political realism, a book analyzing how the political world does work and not very often stating how it ought to work, a book that frequently discloses to a king what calculating and sometimes brutal measures he must carry out to preserve the state and the common good." China Chinese mathematics From the earliest the Chinese used a positional decimal system on counting boards in order to calculate. To express 10, a single rod is placed in the second box from the right. The spoken language uses a similar system to English: e.g. four thousand two hundred seven. No symbol was used for zero. By the 1st century BCE, negative numbers and decimal fractions were in use and The Nine Chapters on the Mathematical Art included methods for extracting higher order roots by Horner's method and solving linear equations and by Pythagoras' theorem. Cubic equations were solved in the Tang dynasty and solutions of equations of order higher than 3 appeared in print in 1245 CE by Ch'in Chiu-shao. Pascal's triangle for binomial coefficients was described around 1100 by Jia Xian. Although the first attempts at an axiomatization of geometry appear in the Mohist canon in 330 BCE, Liu Hui developed algebraic methods in geometry in the 3rd century CE and also calculated pi to 5 significant figures. In 480, Zu Chongzhi improved this by discovering the ratio which remained the most accurate value for 1200 years. Astronomical observations Astronomical observations from China constitute the longest continuous sequence from any civilization and include records of sunspots (112 records from 364 BCE), supernovas (1054), lunar and solar eclipses. By the 12th century, they could reasonably accurately make predictions of eclipses, but the knowledge of this was lost during the Ming dynasty, so that the Jesuit Matteo Ricci gained much favor in 1601 by his predictions. By 635 Chinese astronomers had observed that the tails of comets always point away from the sun. From antiquity, the Chinese used an equatorial system for describing the skies and a star map from 940 was drawn using a cylindrical (Mercator) projection. The use of an armillary sphere is recorded from the 4th century BCE and a sphere permanently mounted in equatorial axis from 52 BCE. In 125 CE Zhang Heng used waterpower to rotate the sphere in real time. This included rings for the meridian and ecliptic. By 1270 they had incorporated the principles of the Arab torquetum. In the Song Empire (960–1279) of Imperial China, Chinese scholar-officials unearthed, studied, and cataloged ancient artifacts. Inventions To better prepare for calamities, Zhang Heng invented a seismometer in 132 CE which provided instant alert to authorities in the capital Luoyang that an earthquake had occurred in a location indicated by a specific cardinal or ordinal direction. Although no tremors could be felt in the capital when Zhang told the court that an earthquake had just occurred in the northwest, a message came soon afterwards that an earthquake had indeed struck northwest of Luoyang (in what is now modern Gansu). Zhang called his device the 'instrument for measuring the seasonal winds and the movements of the Earth' (Houfeng didong yi 候风地动仪), so-named because he and others thought that earthquakes were most likely caused by the enormous compression of trapped air. There are many notable contributors to early Chinese disciplines, inventions, and practices throughout the ages. One of the best examples would be the medieval Song Chinese Shen Kuo (1031–1095), a polymath and statesman who was the first to describe the magnetic-needle compass used for navigation, discovered the concept of true north, improved the design of the astronomical gnomon, armillary sphere, sight tube, and clepsydra, and described the use of drydocks to repair boats. After observing the natural process of the inundation of silt and the find of marine fossils in the Taihang Mountains (hundreds of miles from the Pacific Ocean), Shen Kuo devised a theory of land formation, or geomorphology. He also adopted a theory of gradual climate change in regions over time, after observing petrified bamboo found underground at Yan'an, Shaanxi province. If not for Shen Kuo's writing, the architectural works of Yu Hao would be little known, along with the inventor of movable type printing, Bi Sheng (990–1051). Shen's contemporary Su Song (1020–1101) was also a brilliant polymath, an astronomer who created a celestial atlas of star maps, wrote a treatise related to botany, zoology, mineralogy, and metallurgy, and had erected a large astronomical clocktower in Kaifeng city in 1088. To operate the crowning armillary sphere, his clocktower featured an escapement mechanism and the world's oldest known use of an endless power-transmitting chain drive. The Jesuit China missions of the 16th and 17th centuries "learned to appreciate the scientific achievements of this ancient culture and made them known in Europe. Through their correspondence European scientists first learned about the Chinese science and culture." Western academic thought on the history of Chinese technology and science was galvanized by the work of Joseph Needham and the Needham Research Institute. Among the technological accomplishments of China were, according to the British scholar Needham, early seismological detectors (Zhang Heng in the 2nd century), the water-powered celestial globe (Zhang Heng), matches, the independent invention of the decimal system, dry docks, sliding calipers, the double-action piston pump, cast iron, the blast furnace, the iron plough, the multi-tube seed drill, the wheelbarrow, the suspension bridge, the winnowing machine, the rotary fan, the parachute, natural gas as fuel, the raised-relief map, the propeller, the crossbow, and a solid fuel rocket, the multistage rocket, the horse collar, along with contributions in logic, astronomy, medicine, and other fields. However, cultural factors prevented these Chinese achievements from developing into "modern science". According to Needham, it may have been the religious and philosophical framework of Chinese intellectuals which made them unable to accept the ideas of laws of nature: Classical antiquity The contributions of the Ancient Egyptians and Mesopotamians in the areas of astronomy, mathematics, and medicine had entered and shaped Greek natural philosophy of classical antiquity, whereby formal attempts were made to provide explanations of events in the physical world based on natural causes. Inquiries were also aimed at such practical goals such as establishing a reliable calendar or determining how to cure a variety of illnesses. The ancient people who were considered the first scientists may have thought of themselves as natural philosophers, as practitioners of a skilled profession (for example, physicians), or as followers of a religious tradition (for example, temple healers). Pre-socratics The earliest Greek philosophers, known as the pre-Socratics, provided competing answers to the question found in the myths of their neighbors: "How did the ordered cosmos in which we live come to be?" The pre-Socratic philosopher Thales (640–546 BCE) of Miletus, identified by later authors such as Aristotle as the first of the Ionian philosophers, postulated non-supernatural explanations for natural phenomena. For example, that land floats on water and that earthquakes are caused by the agitation of the water upon which the land floats, rather than the god Poseidon. Thales' student Pythagoras of Samos founded the Pythagorean school, which investigated mathematics for its own sake, and was the first to postulate that the Earth is spherical in shape. Leucippus (5th century BCE) introduced atomism, the theory that all matter is made of indivisible, imperishable units called atoms. This was greatly expanded on by his pupil Democritus and later Epicurus. Natural philosophy Plato and Aristotle produced the first systematic discussions of natural philosophy, which did much to shape later investigations of nature. Their development of deductive reasoning was of particular importance and usefulness to later scientific inquiry. Plato founded the Platonic Academy in 387 BCE, whose motto was "Let none unversed in geometry enter here," and also turned out many notable philosophers. Plato's student Aristotle introduced empiricism and the notion that universal truths can be arrived at via observation and induction, thereby laying the foundations of the scientific method. Aristotle also produced many biological writings that were empirical in nature, focusing on biological causation and the diversity of life. He made countless observations of nature, especially the habits and attributes of plants and animals on Lesbos, classified more than 540 animal species, and dissected at least 50. Aristotle's writings profoundly influenced subsequent Islamic and European scholarship, though they were eventually superseded in the Scientific Revolution. Aristotle also contributed to theories of the elements and the cosmos. He believed that the celestial bodies (such as the planets and the Sun) had something called an unmoved mover that put the celestial bodies in motion. Aristotle tried to explain everything through mathematics and physics, but sometimes explained things such as the motion of celestial bodies through a higher power such as God. Aristotle did not have the technological advancements that would have explained the motion of celestial bodies. In addition, Aristotle had many views on the elements. He believed that everything was derived of the elements earth, water, air, fire, and lastly the Aether. The Aether was a celestial element, and therefore made up the matter of the celestial bodies. The elements of earth, water, air and fire were derived of a combination of two of the characteristics of hot, wet, cold, and dry, and all had their inevitable place and motion. The motion of these elements begins with earth being the closest to "the Earth," then water, air, fire, and finally Aether. In addition to the makeup of all things, Aristotle came up with theories as to why things did not return to their natural motion. He understood that water sits above earth, air above water, and fire above air in their natural state. He explained that although all elements must return to their natural state, the human body and other living things have a constraint on the elements - thus not allowing the elements making one who they are to return to their natural state. The important legacy of this period included substantial advances in factual knowledge, especially in anatomy, zoology, botany, mineralogy, geography, mathematics and astronomy; an awareness of the importance of certain scientific problems, especially those related to the problem of change and its causes; and a recognition of the methodological importance of applying mathematics to natural phenomena and of undertaking empirical research. In the Hellenistic age scholars frequently employed the principles developed in earlier Greek thought: the application of mathematics and deliberate empirical research, in their scientific investigations. Thus, clear unbroken lines of influence lead from ancient Greek and Hellenistic philosophers, to medieval Muslim philosophers and scientists, to the European Renaissance and Enlightenment, to the secular sciences of the modern day. Neither reason nor inquiry began with the Ancient Greeks, but the Socratic method did, along with the idea of Forms, give great advances in geometry, logic, and the natural sciences. According to Benjamin Farrington, former Professor of Classics at Swansea University: "Men were weighing for thousands of years before Archimedes worked out the laws of equilibrium; they must have had practical and intuitional knowledge of the principals involved. What Archimedes did was to sort out the theoretical implications of this practical knowledge and present the resulting body of knowledge as a logically coherent system." and again: "With astonishment we find ourselves on the threshold of modern science. Nor should it be supposed that by some trick of translation the extracts have been given an air of modernity. Far from it. The vocabulary of these writings and their style are the source from which our own vocabulary and style have been derived." Greek astronomy The astronomer Aristarchus of Samos was the first known person to propose a heliocentric model of the Solar System, while the geographer Eratosthenes accurately calculated the circumference of the Earth. Hipparchus (c. 190 – c. 120 BCE) produced the first systematic star catalog. The level of achievement in Hellenistic astronomy and engineering is impressively shown by the Antikythera mechanism (150–100 BCE), an analog computer for calculating the position of planets. Technological artifacts of similar complexity did not reappear until the 14th century, when mechanical astronomical clocks appeared in Europe. Hellenistic medicine There was not a defined societal structure for healthcare during the age of Hippocrates. At that time, society was not organized and knowledgeable as people still relied on pure religious reasoning to explain illnesses. Hippocrates introduced the first healthcare system based on science and clinical protocols. Hippocrates' theories about physics and medicine helped pave the way in creating an organized medical structure for society. In medicine, Hippocrates (c. 460 BC – c. 370 BCE) and his followers were the first to describe many diseases and medical conditions and developed the Hippocratic Oath for physicians, still relevant and in use today. Hippocrates' ideas are expressed in The Hippocratic Corpus. The collection notes descriptions of medical philosophies and how disease and lifestyle choices reflect on the physical body. Hippocrates influenced a Westernized, professional relationship among physician and patient. Hippocrates is also known as "the Father of Medicine".Herophilos (335–280 BCE) was the first to base his conclusions on dissection of the human body and to describe the nervous system. Galen (129 – c. 200 CE) performed many audacious operations—including brain and eye surgeries— that were not tried again for almost two millennia. Greek mathematics In Hellenistic Egypt, the mathematician Euclid laid down the foundations of mathematical rigor and introduced the concepts of definition, axiom, theorem and proof still in use today in his Elements, considered the most influential textbook ever written. Archimedes, considered one of the greatest mathematicians of all time, is credited with using the method of exhaustion to calculate the area under the arc of a parabola with the summation of an infinite series, and gave a remarkably accurate approximation of pi. He is also known in physics for laying the foundations of hydrostatics, statics, and the explanation of the principle of the lever. Other developments Theophrastus wrote some of the earliest descriptions of plants and animals, establishing the first taxonomy and looking at minerals in terms of their properties such as hardness. Pliny the Elder produced what is one of the largest encyclopedias of the natural world in 77 CE and must be regarded as the rightful successor to Theophrastus. For example, he accurately describes the octahedral shape of the diamond and proceeded to mention that diamond dust is used by engravers to cut and polish other gems owing to its great hardness. His recognition of the importance of crystal shape is a precursor to modern crystallography, while mention of numerous other minerals presages mineralogy. He also recognizes that other minerals have characteristic crystal shapes, but in one example, confuses the crystal habit with the work of lapidaries. He was also the first to recognize that amber was a fossilized resin from pine trees because he had seen samples with trapped insects within them. The development of the field of archaeology has its roots with history and with those who were interested in the past, such as kings and queens who wanted to show past glories of their respective nations. The 5th-century-BCE Greek historian Herodotus was the first scholar to systematically study the past and perhaps the first to examine artifacts. Greek scholarship under Roman rule During the rule of Rome, famous historians such as Polybius, Livy and Plutarch documented the rise of the Roman Republic, and the organization and histories of other nations, while statesmen like Julius Caesar, Cicero, and others provided examples of the politics of the republic and Rome's empire and wars. The study of politics during this age was oriented toward understanding history, understanding methods of governing, and describing the operation of governments. The Roman conquest of Greece did not diminish learning and culture in the Greek provinces. On the contrary, the appreciation of Greek achievements in literature, philosophy, politics, and the arts by Rome's upper class coincided with the increased prosperity of the Roman Empire. Greek settlements had existed in Italy for centuries and the ability to read and speak Greek was not uncommon in Italian cities such as Rome. Moreover, the settlement of Greek scholars in Rome, whether voluntarily or as slaves, gave Romans access to teachers of Greek literature and philosophy. Conversely, young Roman scholars also studied abroad in Greece and upon their return to Rome, were able to convey Greek achievements to their Latin leadership. And despite the translation of a few Greek texts into Latin, Roman scholars who aspired to the highest level did so using the Greek language. The Roman statesman and philosopher Cicero (106 – 43 BCE) was a prime example. He had studied under Greek teachers in Rome and then in Athens and Rhodes. He mastered considerable portions of Greek philosophy, wrote Latin treatises on several topics, and even wrote Greek commentaries of Plato's Timaeus as well as a Latin translation of it, which has not survived. In the beginning, support for scholarship in Greek knowledge was almost entirely funded by the Roman upper class. There were all sorts of arrangements, ranging from a talented scholar being attached to a wealthy household to owning educated Greek-speaking slaves. In exchange, scholars who succeeded at the highest level had an obligation to provide advice or intellectual companionship to their Roman benefactors, or to even take care of their libraries. The less fortunate or accomplished ones would teach their children or perform menial tasks. The level of detail and sophistication of Greek knowledge was adjusted to suit the interests of their Roman patrons. That meant popularizing Greek knowledge by presenting information that were of practical value such as medicine or logic (for courts and politics) but excluding subtle details of Greek metaphysics and epistemology. Beyond the basics, the Romans did not value natural philosophy and considered it an amusement for leisure time. Commentaries and encyclopedias were the means by which Greek knowledge was popularized for Roman audiences. The Greek scholar Posidonius (c. 135-c. 51 BCE), a native of Syria, wrote prolifically on history, geography, moral philosophy, and natural philosophy. He greatly influenced Latin writers such as Marcus Terentius Varro (116-27 BCE), who wrote the encyclopedia Nine Books of Disciplines, which covered nine arts: grammar, rhetoric, logic, arithmetic, geometry, astronomy, musical theory, medicine, and architecture. The Disciplines became a model for subsequent Roman encyclopedias and Varro's nine liberal arts were considered suitable education for a Roman gentleman. The first seven of Varro's nine arts would later define the seven liberal arts of medieval schools. The pinnacle of the popularization movement was the Roman scholar Pliny the Elder (23/24–79 CE), a native of northern Italy, who wrote several books on the history of Rome and grammar. His most famous work was his voluminous Natural History. After the death of the Roman Emperor Marcus Aurelius in 180 CE, the favorable conditions for scholarship and learning in the Roman Empire were upended by political unrest, civil war, urban decay, and looming economic crisis. In around 250 CE, barbarians began attacking and invading the Roman frontiers. These combined events led to a general decline in political and economic conditions. The living standards of the Roman upper class was severely impacted, and their loss of leisure diminished scholarly pursuits. Moreover, during the 3rd and 4th centuries CE, the Roman Empire was administratively divided into two halves: Greek East and Latin West. These administrative divisions weakened the intellectual contact between the two regions. Eventually, both halves went their separate ways, with the Greek East becoming the Byzantine Empire. Christianity was also steadily expanding during this time and soon became a major patron of education in the Latin West. Initially, the Christian church adopted some of the reasoning tools of Greek philosophy in the 2nd and 3rd centuries CE to defend its faith against sophisticated opponents. Nevertheless, Greek philosophy received a mixed reception from leaders and adherents of the Christian faith. Some such as Tertullian (c. 155-c. 230 CE) were vehemently opposed to philosophy, denouncing it as heretic. Others such as Augustine of Hippo (354-430 CE) were ambivalent and defended Greek philosophy and science as the best ways to understand the natural world and therefore treated it as a handmaiden (or servant) of religion. Education in the West began its gradual decline, along with the rest of Western Roman Empire, due to invasions by Germanic tribes, civil unrest, and economic collapse. Contact with the classical tradition was lost in specific regions such as Roman Britain and northern Gaul but continued to exist in Rome, northern Italy, southern Gaul, Spain, and North Africa. Middle Ages In the Middle Ages, the classical learning continued in three major linguistic cultures and civilizations: Greek (the Byzantine Empire), Arabic (the Islamic world), and Latin (Western Europe). Byzantine Empire Preservation of Greek heritage The fall of the Western Roman Empire led to a deterioration of the classical tradition in the western part (or Latin West) of Europe in the 400s. In contrast, the Eastern Roman or Byzantine Empire resisted the barbarian attacks and preserved and improved the learning. While the Byzantine Empire still held learning centers such as Constantinople, Alexandria and Antioch, Western Europe's knowledge was concentrated in monasteries until the development of medieval universities in the 12th centuries. The curriculum of monastic schools included the study of the few available ancient texts and of new works on practical subjects like medicine and timekeeping. In the sixth century in the Byzantine Empire, Isidore of Miletus compiled Archimedes' mathematical works in the Archimedes Palimpsest, where all Archimedes' mathematical contributions were collected and studied. John Philoponus, another Byzantine scholar, was the first to question Aristotle's teaching of physics, introducing the theory of impetus. The theory of impetus was an auxiliary or secondary theory of Aristotelian dynamics, put forth initially to explain projectile motion against gravity. It is the intellectual precursor to the concepts of inertia, momentum and acceleration in classical mechanics. The works of John Philoponus inspired Galileo Galilei ten centuries later. Collapse During the Fall of Constantinople in 1453, a number of Greek scholars fled to North Italy in which they fueled the era later commonly known as the "Renaissance" as they brought with them a great deal of classical learning including an understanding of botany, medicine, and zoology. Byzantium also gave the West important inputs: John Philoponus' criticism of Aristotelian physics, and the works of Dioscorides. Islamic world This was the period (8th–14th century CE) of the Islamic Golden Age where commerce thrived, and new ideas and technologies emerged such as the importation of papermaking from China, which made the copying of manuscripts inexpensive. Translations and Hellenization The eastward transmission of Greek heritage to Western Asia was a slow and gradual process that spanned over a thousand years, beginning with the Asian conquests of Alexander the Great in 335 BCE to the founding of Islam in the 7th century CE. The birth and expansion of Islam during the 7th century was quickly followed by its Hellenization. Knowledge of Greek conceptions of the world was preserved and absorbed into Islamic theology, law, culture, and commerce, which were aided by the translations of traditional Greek texts and some Syriac intermediary sources into Arabic during the 8th–9th century. Education and scholarly pursuits Madrasas were centers for many different religious and scientific studies and were the culmination of different institutions such as mosques based around religious studies, housing for out-of-town visitors, and finally educational institutions focused on the natural sciences. Unlike Western universities, students at a madrasa would learn from one specific teacher, who would issue a certificate at the completion of their studies called an Ijazah. An Ijazah differs from a western university degree in many ways one being that it is issued by a single person rather than an institution, and another being that it is not an individual degree declaring adequate knowledge over broad subjects, but rather a license to teach and pass on a very specific set of texts. Women were also allowed to attend madrasas, as both students and teachers, something not seen in high western education until the 1800s. Madrasas were more than just academic centers. The Suleymaniye Mosque, for example, was one of the earliest and most well-known madrasas, which was built by Suleiman the Magnificent in the 16th century The Suleymaniye Mosque was home to a hospital and medical college, a kitchen, and children's school, as well as serving as a temporary home for travelers. Higher education at a madrasa (or college) was focused on Islamic law and religious science and students had to engage in self-study for everything else. And despite the occasional theological backlash, many Islamic scholars of science were able to conduct their work in relatively tolerant urban centers (e.g., Baghdad and Cairo) and were protected by powerful patrons. They could also travel freely and exchange ideas as there were no political barriers within the unified Islamic state. Islamic science during this time was primarily focused on the correction, extension, articulation, and application of Greek ideas to new problems. Advancements in mathematics Most of the achievements by Islamic scholars during this period were in mathematics. Arabic mathematics was a direct descendant of Greek and Indian mathematics. For instance, what is now known as Arabic numerals originally came from India, but Muslim mathematicians made several key refinements to the number system, such as the introduction of decimal point notation. Mathematicians such as Muhammad ibn Musa al-Khwarizmi (c. 780–850) gave his name to the concept of the algorithm, while the term algebra is derived from al-jabr, the beginning of the title of one of his publications. Islamic trigonometry continued from the works of Ptolemy's Almagest and Indian Siddhanta, from which they added trigonometric functions, drew up tables, and applied trignometry to spheres and planes. Many of their engineers, instruments makers, and surveyors contributed books in applied mathematics. It was in astronomy where Islamic mathematicians made their greatest contributions. Al-Battani (c. 858–929) improved the measurements of Hipparchus, preserved in the translation of Ptolemy's Hè Megalè Syntaxis (The great treatise) translated as Almagest. Al-Battani also improved the precision of the measurement of the precession of the Earth's axis. Corrections were made to Ptolemy's geocentric model by al-Battani, Ibn al-Haytham, Averroes and the Maragha astronomers such as Nasir al-Din al-Tusi, Mu'ayyad al-Din al-Urdi and Ibn al-Shatir. Scholars with geometric skills made significant improvements to the earlier classical texts on light and sight by Euclid, Aristotle, and Ptolemy. The earliest surviving Arabic treatises were written in the 9th century by Abū Ishāq al-Kindī, Qustā ibn Lūqā, and (in fragmentary form) Ahmad ibn Isā. Later in the 11th century, Ibn al-Haytham (known as Alhazen in the West), a mathematician and astronomer, synthesized a new theory of vision based on the works of his predecessors. His new theory included a complete system of geometrical optics, which was set in great detail in his Book of Optics. His book was translated into Latin and was relied upon as a principal source on the science of optics in Europe until the 17th century. Institutionalization of medicine The medical sciences were prominently cultivated in the Islamic world. The works of Greek medical theories, especially those of Galen, were translated into Arabic and there was an outpouring of medical texts by Islamic physicians, which were aimed at organizing, elaborating, and disseminating classical medical knowledge. Medical specialties started to emerge, such as those involved in the treatment of eye diseases such as cataracts. Ibn Sina (known as Avicenna in the West, c. 980–1037) was a prolific Persian medical encyclopedist wrote extensively on medicine, with his two most notable works in medicine being the Kitāb al-shifāʾ ("Book of Healing") and The Canon of Medicine, both of which were used as standard medicinal texts in both the Muslim world and in Europe well into the 17th century. Amongst his many contributions are the discovery of the contagious nature of infectious diseases, and the introduction of clinical pharmacology. Institutionalization of medicine was another important achievement in the Islamic world. Although hospitals as an institution for the sick emerged in the Byzantium empire, the model of institutionalized medicine for all social classes was extensive in the Islamic empire and was scattered throughout. In addition to treating patients, physicians could teach apprentice physicians, as well write and do research. The discovery of the pulmonary transit of blood in the human body by Ibn al-Nafis occurred in a hospital setting. Decline Islamic science began its decline in the 12th–13th century, before the Renaissance in Europe, due in part to the Christian reconquest of Spain and the Mongol conquests in the East in the 11th–13th century. The Mongols sacked Baghdad, capital of the Abbasid caliphate, in 1258, which ended the Abbasid empire. Nevertheless, many of the conquerors became patrons of the sciences. Hulagu Khan, for example, who led the siege of Baghdad, became a patron of the Maragheh observatory. Islamic astronomy continued to flourish into the 16th century. Western Europe By the eleventh century, most of Europe had become Christian; stronger monarchies emerged; borders were restored; technological developments and agricultural innovations were made, increasing the food supply and population. Classical Greek texts were translated from Arabic and Greek into Latin, stimulating scientific discussion in Western Europe. In classical antiquity, Greek and Roman taboos had meant that dissection was usually banned, but in the Middle Ages medical teachers and students at Bologna began to open human bodies, and Mondino de Luzzi (–1326) produced the first known anatomy textbook based on human dissection. As a result of the Pax Mongolica, Europeans, such as Marco Polo, began to venture further and further east. The written accounts of Polo and his fellow travelers inspired other Western European maritime explorers to search for a direct sea route to Asia, ultimately leading to the Age of Discovery. Technological advances were also made, such as the early flight of Eilmer of Malmesbury (who had studied mathematics in 11th-century England), and the metallurgical achievements of the Cistercian blast furnace at Laskill. Medieval universities An intellectual revitalization of Western Europe started with the birth of medieval universities in the 12th century. These urban institutions grew from the informal scholarly activities of learned friars who visited monasteries, consulted libraries, and conversed with other fellow scholars. A friar who became well-known would attract a following of disciples, giving rise to a brotherhood of scholars (or collegium in Latin). A collegium might travel to a town or request a monastery to host them. However, if the number of scholars within a collegium grew too large, they would opt to settle in a town instead. As the number of collegia within a town grew, the collegia might request that their king grant them a charter that would convert them into a universitas. Many universities were chartered during this period, with the first in Bologna in 1088, followed by Paris in 1150, Oxford in 1167, and Cambridge in 1231. The granting of a charter meant that the medieval universities were partially sovereign and independent from local authorities. Their independence allowed them to conduct themselves and judge their own members based on their own rules. Furthermore, as initially religious institutions, their faculties and students were protected from capital punishment (e.g., gallows). Such independence was a matter of custom, which could, in principle, be revoked by their respective rulers if they felt threatened. Discussions of various subjects or claims at these medieval institutions, no matter how controversial, were done in a formalized way so as to declare such discussions as being within the bounds of a university and therefore protected by the privileges of that institution's sovereignty. A claim could be described as ex cathedra (literally "from the chair", used within the context of teaching) or ex hypothesi (by hypothesis). This meant that the discussions were presented as purely an intellectual exercise that did not require those involved to commit themselves to the truth of a claim or to proselytize. Modern academic concepts and practices such as academic freedom or freedom of inquiry are remnants of these medieval privileges that were tolerated in the past. The curriculum of these medieval institutions centered on the seven liberal arts, which were aimed at providing beginning students with the skills for reasoning and scholarly language. Students would begin their studies starting with the first three liberal arts or Trivium (grammar, rhetoric, and logic) followed by the next four liberal arts or Quadrivium (arithmetic, geometry, astronomy, and music). Those who completed these requirements and received their baccalaureate (or Bachelor of Arts) had the option to join the higher faculty (law, medicine, or theology), which would confer an LLD for a lawyer, an MD for a physician, or ThD for a theologian. Students who chose to remain in the lower faculty (arts) could work towards a Magister (or Master's) degree and would study three philosophies: metaphysics, ethics, and natural philosophy. Latin translations of Aristotle's works such as De Anima (On the Soul) and the commentaries on them were required readings. As time passed, the lower faculty was allowed to confer its own doctoral degree called the PhD. Many of the Masters were drawn to encyclopedias and had used them as textbooks. But these scholars yearned for the complete original texts of the Ancient Greek philosophers, mathematicians, and physicians such as Aristotle, Euclid, and Galen, which were not available to them at the time. These Ancient Greek texts were to be found in the Byzantine Empire and the Islamic World. Translations of Greek and Arabic sources Contact with the Byzantine Empire, and with the Islamic world during the Reconquista and the Crusades, allowed Latin Europe access to scientific Greek and Arabic texts, including the works of Aristotle, Ptolemy, Isidore of Miletus, John Philoponus, Jābir ibn Hayyān, al-Khwarizmi, Alhazen, Avicenna, and Averroes. European scholars had access to the translation programs of Raymond of Toledo, who sponsored the 12th century Toledo School of Translators from Arabic to Latin. Later translators like Michael Scotus would learn Arabic in order to study these texts directly. The European universities aided materially in the translation and propagation of these texts and started a new infrastructure which was needed for scientific communities. In fact, European university put many works about the natural world and the study of nature at the center of its curriculum, with the result that the "medieval university laid far greater emphasis on science than does its modern counterpart and descendent." At the beginning of the 13th century, there were reasonably accurate Latin translations of the main works of almost all the intellectually crucial ancient authors, allowing a sound transfer of scientific ideas via both the universities and the monasteries. By then, the natural philosophy in these texts began to be extended by scholastics such as Robert Grosseteste, Roger Bacon, Albertus Magnus and Duns Scotus. Precursors of the modern scientific method, influenced by earlier contributions of the Islamic world, can be seen already in Grosseteste's emphasis on mathematics as a way to understand nature, and in the empirical approach admired by Bacon, particularly in his Opus Majus. Pierre Duhem's thesis is that Stephen Tempier – the Bishop of Paris – Condemnation of 1277 led to the study of medieval science as a serious discipline, "but no one in the field any longer endorses his view that modern science started in 1277". However, many scholars agree with Duhem's view that the mid-late Middle Ages saw important scientific developments. Medieval science The first half of the 14th century saw much important scientific work, largely within the framework of scholastic commentaries on Aristotle's scientific writings. William of Ockham emphasized the principle of parsimony: natural philosophers should not postulate unnecessary entities, so that motion is not a distinct thing but is only the moving object and an intermediary "sensible species" is not needed to transmit an image of an object to the eye. Scholars such as Jean Buridan and Nicole Oresme started to reinterpret elements of Aristotle's mechanics. In particular, Buridan developed the theory that impetus was the cause of the motion of projectiles, which was a first step towards the modern concept of inertia. The Oxford Calculators began to mathematically analyze the kinematics of motion, making this analysis without considering the causes of motion. In 1348, the Black Death and other disasters sealed a sudden end to philosophic and scientific development. Yet, the rediscovery of ancient texts was stimulated by the Fall of Constantinople in 1453, when many Byzantine scholars sought refuge in the West. Meanwhile, the introduction of printing was to have great effect on European society. The facilitated dissemination of the printed word democratized learning and allowed ideas such as algebra to propagate more rapidly. These developments paved the way for the Scientific Revolution, where scientific inquiry, halted at the start of the Black Death, resumed. Renaissance Revival of learning The renewal of learning in Europe began with 12th century Scholasticism. The Northern Renaissance showed a decisive shift in focus from Aristotelian natural philosophy to chemistry and the biological sciences (botany, anatomy, and medicine). Thus modern science in Europe was resumed in a period of great upheaval: the Protestant Reformation and Catholic Counter-Reformation; the discovery of the Americas by Christopher Columbus; the Fall of Constantinople; but also the re-discovery of Aristotle during the Scholastic period presaged large social and political changes. Thus, a suitable environment was created in which it became possible to question scientific doctrine, in much the same way that Martin Luther and John Calvin questioned religious doctrine. The works of Ptolemy (astronomy) and Galen (medicine) were found not always to match everyday observations. Work by Vesalius on human cadavers found problems with the Galenic view of anatomy. Theophrastus' work on rocks, Peri lithōn, remained authoritative for millennia: its interpretation of fossils was not overturned until after the Scientific Revolution. During the Italian Renaissance, Niccolò Machiavelli established the emphasis of modern political science on direct empirical observation of political institutions and actors. Later, the expansion of the scientific paradigm during the Enlightenment further pushed the study of politics beyond normative determinations. In particular, the study of statistics, to study the subjects of the state, has been applied to polling and voting. In archaeology, the 15th and 16th centuries saw the rise of antiquarians in Renaissance Europe who were interested in the collection of artifacts. Scientific Revolution and birth of New Science The early modern period is seen as a flowering of the European Renaissance. There was a willingness to question previously held truths and search for new answers resulted in a period of major scientific advancements, now known as the Scientific Revolution, which led to the emergence of a New Science that was more mechanistic in its worldview, more integrated with mathematics, and more reliable and open as its knowledge was based on a newly defined scientific method. The Scientific Revolution is a convenient boundary between ancient thought and classical physics, and is traditionally held by most historians to have begun in 1543, when the books De humani corporis fabrica (On the Workings of the Human Body) by Andreas Vesalius, and also De Revolutionibus, by the astronomer Nicolaus Copernicus, were first printed. The period culminated with the publication of the Philosophiæ Naturalis Principia Mathematica in 1687 by Isaac Newton, representative of the unprecedented growth of scientific publications throughout Europe. Other significant scientific advances were made during this time by Galileo Galilei, Edmond Halley, William Harvey, Pierre Fermat, Robert Hooke, Christiaan Huygens, Tycho Brahe, Johannes Kepler, Marin Mersenne, Gottfried Leibniz, Isaac Newton, and Blaise Pascal. In philosophy, major contributions were made by Francis Bacon, Sir Thomas Browne, René Descartes, Baruch Spinoza, Pierre Gassendi, Robert Boyle, and Thomas Hobbes. Christiaan Huygens derived the centripetal and centrifugal forces and was the first to transfer mathematical inquiry to describe unobservable physical phenomena. William Gilbert did some of the earliest experiments with electricity and magnetism, establishing that the Earth itself is magnetic. Heliocentrism The heliocentric model was revived by Nicolaus Copernicus. The theory of Copernicus was that the Earth and all heavenly spheres, containing the planets and other objects in the cosmos, rotated around the Sun. His heliocentric model also proposed that all fixed stars were not rotating on an axis, nor in any motion at all. His theory proposed the yearly rotation of the Earth and the other heavenly spheres around the Sun and was able to calculate the distances of planets using deferents and epicycles. Although these calculations were not completely accurate, Copernicus was able to understand the distance order of each heavenly sphere. The Copernican heliocentric system was a revival of the hypotheses of Aristarchus of Samos and Seleucus of Seleucia. Aristarchus of Samos did propose that the Earth rotated around the Sun but did not mention anything about the other heavenly spheres' order, motion, or rotation. Seleucus of Seleucia also proposed the rotation of the Earth around the Sun but did not mention anything about the other heavenly spheres. In addition, Seleucus of Seleucia understood that the Moon rotated around the Earth and could be used to explain the tides of the oceans, thus further proving his understanding of the heliocentric idea. Newly defined scientific method The scientific method was also better developed as the modern way of thinking emphasized experimentation and reason over traditional considerations. Galileo ("Father of Modern Physics") also made use of experiments to validate physical theories, a key element of the scientific method. Age of Enlightenment Continuation of Scientific Revolution The Scientific Revolution continued into the Age of Enlightenment, which accelerated the development of modern science. Planets and orbits The heliocentric model revived by Nicolaus Copernicus was followed by the model of planetary motion given by Johannes Kepler in the early 17th century, which proposed that the planets follow elliptical orbits, with the Sun at one focus of the ellipse. Calculus and Newtonian mechanics In 1687, Isaac Newton published the Principia Mathematica, detailing two comprehensive and successful physical theories: Newton's laws of motion, which led to classical mechanics; and Newton's law of universal gravitation, which describes the fundamental force of gravity. Emergence of chemistry A decisive moment came when "chemistry" was distinguished from alchemy by Robert Boyle in his work The Sceptical Chymist, in 1661; although the alchemical tradition continued for some time after his work. Other important steps included the gravimetric experimental practices of medical chemists like William Cullen, Joseph Black, Torbern Bergman and Pierre Macquer and through the work of Antoine Lavoisier ("father of modern chemistry") on oxygen and the law of conservation of mass, which refuted phlogiston theory. Modern chemistry emerged from the sixteenth through the eighteenth centuries through the material practices and theories promoted by alchemy, medicine, manufacturing and mining. Circulatory system William Harvey published De Motu Cordis in 1628, which revealed his conclusions based on his extensive studies of vertebrate circulatory systems. He identified the central role of the heart, arteries, and veins in producing blood movement in a circuit, and failed to find any confirmation of Galen's pre-existing notions of heating and cooling functions. The history of early modern biology and medicine is often told through the search for the seat of the soul. Galen in his descriptions of his foundational work in medicine presents the distinctions between arteries, veins, and nerves using the vocabulary of the soul. Scientific societies and journals A critical innovation was the creation of permanent scientific societies and their scholarly journals, which dramatically sped the diffusion of new ideas. Typical was the founding of the Royal Society in London in 1660 and its journal in 1665 the Philosophical Transaction of the Royal Society, the first scientific journal in English. 1665 also saw the first journal in French, the Journal des sçavans. Science drawing on the works of Newton, Descartes, Pascal and Leibniz, science was on a path to modern mathematics, physics and technology by the time of the generation of Benjamin Franklin (1706–1790), Leonhard Euler (1707–1783), Mikhail Lomonosov (1711–1765) and Jean le Rond d'Alembert (1717–1783). Denis Diderot's Encyclopédie, published between 1751 and 1772 brought this new understanding to a wider audience. The impact of this process was not limited to science and technology, but affected philosophy (Immanuel Kant, David Hume), religion (the increasingly significant impact of science upon religion), and society and politics in general (Adam Smith, Voltaire). Developments in geology Geology did not undergo systematic restructuring during the Scientific Revolution but instead existed as a cloud of isolated, disconnected ideas about rocks, minerals, and landforms long before it became a coherent science. Robert Hooke formulated a theory of earthquakes, and Nicholas Steno developed the theory of superposition and argued that fossils were the remains of once-living creatures. Beginning with Thomas Burnet's Sacred Theory of the Earth in 1681, natural philosophers began to explore the idea that the Earth had changed over time. Burnet and his contemporaries interpreted Earth's past in terms of events described in the Bible, but their work laid the intellectual foundations for secular interpretations of Earth history. Post-Scientific Revolution Bioelectricity During the late 18th century, researchers such as Hugh Williamson and John Walsh experimented on the effects of electricity on the human body. Further studies by Luigi Galvani and Alessandro Volta established the electrical nature of what Volta called galvanism. Developments in geology Modern geology, like modern chemistry, gradually evolved during the 18th and early 19th centuries. Benoît de Maillet and the Comte de Buffon saw the Earth as much older than the 6,000 years envisioned by biblical scholars. Jean-Étienne Guettard and Nicolas Desmarest hiked central France and recorded their observations on some of the first geological maps. Aided by chemical experimentation, naturalists such as Scotland's John Walker, Sweden's Torbern Bergman, and Germany's Abraham Werner created comprehensive classification systems for rocks and minerals—a collective achievement that transformed geology into a cutting edge field by the end of the eighteenth century. These early geologists also proposed a generalized interpretations of Earth history that led James Hutton, Georges Cuvier and Alexandre Brongniart, following in the steps of Steno, to argue that layers of rock could be dated by the fossils they contained: a principle first applied to the geology of the Paris Basin. The use of index fossils became a powerful tool for making geological maps, because it allowed geologists to correlate the rocks in one locality with those of similar age in other, distant localities. Birth of modern economics The basis for classical economics forms Adam Smith's An Inquiry into the Nature and Causes of the Wealth of Nations, published in 1776. Smith criticized mercantilism, advocating a system of free trade with division of labour. He postulated an "invisible hand" that regulated economic systems made up of actors guided only by self-interest. The "invisible hand" mentioned in a lost page in the middle of a chapter in the middle of the "Wealth of Nations", 1776, advances as Smith's central message. Social science Anthropology can best be understood as an outgrowth of the Age of Enlightenment. It was during this period that Europeans attempted systematically to study human behavior. Traditions of jurisprudence, history, philology and sociology developed during this time and informed the development of the social sciences of which anthropology was a part. 19th century The 19th century saw the birth of science as a profession. William Whewell had coined the term scientist in 1833, which soon replaced the older term natural philosopher. Developments in physics In physics, the behavior of electricity and magnetism was studied by Giovanni Aldini, Alessandro Volta, Michael Faraday, Georg Ohm, and others. The experiments, theories and discoveries of Michael Faraday, Andre-Marie Ampere, James Clerk Maxwell, and their contemporaries led to the unification of the two phenomena into a single theory of electromagnetism as described by Maxwell's equations. Thermodynamics led to an understanding of heat and the notion of energy being defined. Discovery of Neptune In astronomy, the planet Neptune was discovered. Advances in astronomy and in optical systems in the 19th century resulted in the first observation of an asteroid (1 Ceres) in 1801, and the discovery of Neptune in 1846. Developments in mathematics In mathematics, the notion of complex numbers finally matured and led to a subsequent analytical theory; they also began the use of hypercomplex numbers. Karl Weierstrass and others carried out the arithmetization of analysis for functions of real and complex variables. It also saw rise to new progress in geometry beyond those classical theories of Euclid, after a period of nearly two thousand years. The mathematical science of logic likewise had revolutionary breakthroughs after a similarly long period of stagnation. But the most important step in science at this time were the ideas formulated by the creators of electrical science. Their work changed the face of physics and made possible for new technology to come about such as electric power, electrical telegraphy, the telephone, and radio. Developments in chemistry In chemistry, Dmitri Mendeleev, following the atomic theory of John Dalton, created the first periodic table of elements. Other highlights include the discoveries unveiling the nature of atomic structure and matter, simultaneously with chemistry – and of new kinds of radiation. The theory that all matter is made of atoms, which are the smallest constituents of matter that cannot be broken down without losing the basic chemical and physical properties of that matter, was provided by John Dalton in 1803, although the question took a hundred years to settle as proven. Dalton also formulated the law of mass relationships. In 1869, Dmitri Mendeleev composed his periodic table of elements on the basis of Dalton's discoveries. The synthesis of urea by Friedrich Wöhler opened a new research field, organic chemistry, and by the end of the 19th century, scientists were able to synthesize hundreds of organic compounds. The later part of the 19th century saw the exploitation of the Earth's petrochemicals, after the exhaustion of the oil supply from whaling. By the 20th century, systematic production of refined materials provided a ready supply of products which provided not only energy, but also synthetic materials for clothing, medicine, and everyday disposable resources. Application of the techniques of organic chemistry to living organisms resulted in physiological chemistry, the precursor to biochemistry. Age of the Earth Over the first half of the 19th century, geologists such as Charles Lyell, Adam Sedgwick, and Roderick Murchison applied the new technique to rocks throughout Europe and eastern North America, setting the stage for more detailed, government-funded mapping projects in later decades. Midway through the 19th century, the focus of geology shifted from description and classification to attempts to understand how the surface of the Earth had changed. The first comprehensive theories of mountain building were proposed during this period, as were the first modern theories of earthquakes and volcanoes. Louis Agassiz and others established the reality of continent-covering ice ages, and "fluvialists" like Andrew Crombie Ramsay argued that river valleys were formed, over millions of years by the rivers that flow through them. After the discovery of radioactivity, radiometric dating methods were developed, starting in the 20th century. Alfred Wegener's theory of "continental drift" was widely dismissed when he proposed it in the 1910s, but new data gathered in the 1950s and 1960s led to the theory of plate tectonics, which provided a plausible mechanism for it. Plate tectonics also provided a unified explanation for a wide range of seemingly unrelated geological phenomena. Since 1970 it has served as the unifying principle in geology. Evolution and inheritance Perhaps the most prominent, controversial, and far-reaching theory in all of science has been the theory of evolution by natural selection, which was independently formulated by Charles Darwin and Alfred Wallace. It was described in detail in Darwin's book The Origin of Species, which was published in 1859. In it, Darwin proposed that the features of all living things, including humans, were shaped by natural processes over long periods of time. The theory of evolution in its current form affects almost all areas of biology. Implications of evolution on fields outside of pure science have led to both opposition and support from different parts of society, and profoundly influenced the popular understanding of "man's place in the universe". Separately, Gregor Mendel formulated in the principles of inheritance in 1866, which became the basis of modern genetics. Germ theory Another important landmark in medicine and biology were the successful efforts to prove the germ theory of disease. Following this, Louis Pasteur made the first vaccine against rabies, and also made many discoveries in the field of chemistry, including the asymmetry of crystals. In 1847, Hungarian physician Ignác Fülöp Semmelweis dramatically reduced the occurrence of puerperal fever by simply requiring physicians to wash their hands before attending to women in childbirth. This discovery predated the germ theory of disease. However, Semmelweis' findings were not appreciated by his contemporaries and handwashing came into use only with discoveries by British surgeon Joseph Lister, who in 1865 proved the principles of antisepsis. Lister's work was based on the important findings by French biologist Louis Pasteur. Pasteur was able to link microorganisms with disease, revolutionizing medicine. He also devised one of the most important methods in preventive medicine, when in 1880 he produced a vaccine against rabies. Pasteur invented the process of pasteurization, to help prevent the spread of disease through milk and other foods. Schools of economics Karl Marx developed an alternative economic theory, called Marxian economics. Marxian economics is based on the labor theory of value and assumes the value of good to be based on the amount of labor required to produce it. Under this axiom, capitalism was based on employers not paying the full value of workers labor to create profit. The Austrian School responded to Marxian economics by viewing entrepreneurship as driving force of economic development. This replaced the labor theory of value by a system of supply and demand. Founding of psychology Psychology as a scientific enterprise that was independent from philosophy began in 1879 when Wilhelm Wundt founded the first laboratory dedicated exclusively to psychological research (in Leipzig). Other important early contributors to the field include Hermann Ebbinghaus (a pioneer in memory studies), Ivan Pavlov (who discovered classical conditioning), William James, and Sigmund Freud. Freud's influence has been enormous, though more as cultural icon than a force in scientific psychology. Modern sociology Modern sociology emerged in the early 19th century as the academic response to the modernization of the world. Among many early sociologists (e.g., Émile Durkheim), the aim of sociology was in structuralism, understanding the cohesion of social groups, and developing an "antidote" to social disintegration. Max Weber was concerned with the modernization of society through the concept of rationalization, which he believed would trap individuals in an "iron cage" of rational thought. Some sociologists, including Georg Simmel and W. E. B. Du Bois, used more microsociological, qualitative analyses. This microlevel approach played an important role in American sociology, with the theories of George Herbert Mead and his student Herbert Blumer resulting in the creation of the symbolic interactionism approach to sociology. In particular, just Auguste Comte, illustrated with his work the transition from a theological to a metaphysical stage and, from this, to a positive stage. Comte took care of the classification of the sciences as well as a transit of humanity towards a situation of progress attributable to a re-examination of nature according to the affirmation of 'sociality' as the basis of the scientifically interpreted society. Romanticism The Romantic Movement of the early 19th century reshaped science by opening up new pursuits unexpected in the classical approaches of the Enlightenment. The decline of Romanticism occurred because a new movement, Positivism, began to take hold of the ideals of the intellectuals after 1840 and lasted until about 1880. At the same time, the romantic reaction to the Enlightenment produced thinkers such as Johann Gottfried Herder and later Wilhelm Dilthey whose work formed the basis for the culture concept which is central to the discipline. Traditionally, much of the history of the subject was based on colonial encounters between Western Europe and the rest of the world, and much of 18th- and 19th-century anthropology is now classed as scientific racism. During the late 19th century, battles over the "study of man" took place between those of an "anthropological" persuasion (relying on anthropometrical techniques) and those of an "ethnological" persuasion (looking at cultures and traditions), and these distinctions became part of the later divide between physical anthropology and cultural anthropology, the latter ushered in by the students of Franz Boas. 20th century Science advanced dramatically during the 20th century. There were new and radical developments in the physical and life sciences, building on the progress from the 19th century. Theory of relativity and quantum mechanics The beginning of the 20th century brought the start of a revolution in physics. The long-held theories of Newton were shown not to be correct in all circumstances. Beginning in 1900, Max Planck, Albert Einstein, Niels Bohr and others developed quantum theories to explain various anomalous experimental results, by introducing discrete energy levels. Not only did quantum mechanics show that the laws of motion did not hold on small scales, but the theory of general relativity, proposed by Einstein in 1915, showed that the fixed background of spacetime, on which both Newtonian mechanics and special relativity depended, could not exist. In 1925, Werner Heisenberg and Erwin Schrödinger formulated quantum mechanics, which explained the preceding quantum theories. The observation by Edwin Hubble in 1929 that the speed at which galaxies recede positively correlates with their distance, led to the understanding that the universe is expanding, and the formulation of the Big Bang theory by Georges Lemaître. Currently, general relativity and quantum mechanics are inconsistent with each other, and efforts are underway to unify the two. Big science In 1938 Otto Hahn and Fritz Strassmann discovered nuclear fission with radiochemical methods, and in 1939 Lise Meitner and Otto Robert Frisch wrote the first theoretical interpretation of the fission process, which was later improved by Niels Bohr and John A. Wheeler. Further developments took place during World War II, which led to the practical application of radar and the development and use of the atomic bomb. Around this time, Chien-Shiung Wu was recruited by the Manhattan Project to help develop a process for separating uranium metal into U-235 and U-238 isotopes by Gaseous diffusion. She was an expert experimentalist in beta decay and weak interaction physics. Wu designed an experiment (see Wu experiment) that enabled theoretical physicists Tsung-Dao Lee and Chen-Ning Yang to disprove the law of parity experimentally, winning them a Nobel Prize in 1957. Though the process had begun with the invention of the cyclotron by Ernest O. Lawrence in the 1930s, physics in the postwar period entered into a phase of what historians have called "Big Science", requiring massive machines, budgets, and laboratories in order to test their theories and move into new frontiers. The primary patron of physics became state governments, who recognized that the support of "basic" research could often lead to technologies useful to both military and industrial applications. Big Bang George Gamow, Ralph Alpher, and Robert Herman had calculated that there should be evidence for a Big Bang in the background temperature of the universe. In 1964, Arno Penzias and Robert Wilson discovered a 3 Kelvin background hiss in their Bell Labs radiotelescope (the Holmdel Horn Antenna), which was evidence for this hypothesis, and formed the basis for a number of results that helped determine the age of the universe. Space exploration In 1925, Cecilia Payne-Gaposchkin determined that stars were composed mostly of hydrogen and helium. She was dissuaded by astronomer Henry Norris Russell from publishing this finding in her PhD thesis because of the widely held belief that stars had the same composition as the Earth. However, four years later, in 1929, Henry Norris Russell came to the same conclusion through different reasoning and the discovery was eventually accepted. In 1987, supernova SN 1987A was observed by astronomers on Earth both visually, and in a triumph for neutrino astronomy, by the solar neutrino detectors at Kamiokande. But the solar neutrino flux was a fraction of its theoretically expected value. This discrepancy forced a change in some values in the standard model for particle physics. Advances in genetics In the early 20th century, the study of heredity became a major investigation after the rediscovery in 1900 of the laws of inheritance developed by Mendel. The 20th century also saw the integration of physics and chemistry, with chemical properties explained as the result of the electronic structure of the atom. Linus Pauling's book on The Nature of the Chemical Bond used the principles of quantum mechanics to deduce bond angles in ever-more complicated molecules. Pauling's work culminated in the physical modelling of DNA, the secret of life (in the words of Francis Crick, 1953). In the same year, the Miller–Urey experiment demonstrated in a simulation of primordial processes, that basic constituents of proteins, simple amino acids, could themselves be built up from simpler molecules, kickstarting decades of research into the chemical origins of life. By 1953, James D. Watson and Francis Crick clarified the basic structure of DNA, the genetic material for expressing life in all its forms, building on the work of Maurice Wilkins and Rosalind Franklin, suggested that the structure of DNA was a double helix. In their famous paper "Molecular structure of Nucleic Acids" In the late 20th century, the possibilities of genetic engineering became practical for the first time, and a massive international effort began in 1990 to map out an entire human genome (the Human Genome Project). The discipline of ecology typically traces its origin to the synthesis of Darwinian evolution and Humboldtian biogeography, in the late 19th and early 20th centuries. Equally important in the rise of ecology, however, were microbiology and soil science—particularly the cycle of life concept, prominent in the work Louis Pasteur and Ferdinand Cohn. The word ecology was coined by Ernst Haeckel, whose particularly holistic view of nature in general (and Darwin's theory in particular) was important in the spread of ecological thinking. In the 1930s, Arthur Tansley and others began developing the field of ecosystem ecology, which combined experimental soil science with physiological concepts of energy and the techniques of field biology. Neuroscience as a distinct discipline The understanding of neurons and the nervous system became increasingly precise and molecular during the 20th century. For example, in 1952, Alan Lloyd Hodgkin and Andrew Huxley presented a mathematical model for transmission of electrical signals in neurons of the giant axon of a squid, which they called "action potentials", and how they are initiated and propagated, known as the Hodgkin–Huxley model. In 1961–1962, Richard FitzHugh and J. Nagumo simplified Hodgkin–Huxley, in what is called the FitzHugh–Nagumo model. In 1962, Bernard Katz modeled neurotransmission across the space between neurons known as synapses. Beginning in 1966, Eric Kandel and collaborators examined biochemical changes in neurons associated with learning and memory storage in Aplysia. In 1981 Catherine Morris and Harold Lecar combined these models in the Morris–Lecar model. Such increasingly quantitative work gave rise to numerous biological neuron models and models of neural computation. Neuroscience began to be recognized as a distinct academic discipline in its own right. Eric Kandel and collaborators have cited David Rioch, Francis O. Schmitt, and Stephen Kuffler as having played critical roles in establishing the field. Plate tectonics Geologists' embrace of plate tectonics became part of a broadening of the field from a study of rocks into a study of the Earth as a planet. Other elements of this transformation include: geophysical studies of the interior of the Earth, the grouping of geology with meteorology and oceanography as one of the "earth sciences", and comparisons of Earth and the solar system's other rocky planets. Applications In terms of applications, a massive number of new technologies were developed in the 20th century. Technologies such as electricity, the incandescent light bulb, the automobile and the phonograph, first developed at the end of the 19th century, were perfected and universally deployed. The first car was introduced by Karl Benz in 1885. The first airplane flight occurred in 1903, and by the end of the century airliners flew thousands of miles in a matter of hours. The development of the radio, television and computers caused massive changes in the dissemination of information. Advances in biology also led to large increases in food production, as well as the elimination of diseases such as polio by Dr. Jonas Salk. Gene mapping and gene sequencing, invented by Drs. Mark Skolnik and Walter Gilbert, respectively, are the two technologies that made the Human Genome Project feasible. Computer science, built upon a foundation of theoretical linguistics, discrete mathematics, and electrical engineering, studies the nature and limits of computation. Subfields include computability, computational complexity, database design, computer networking, artificial intelligence, and the design of computer hardware. One area in which advances in computing have contributed to more general scientific development is by facilitating large-scale archiving of scientific data. Contemporary computer science typically distinguishes itself by emphasizing mathematical 'theory' in contrast to the practical emphasis of software engineering. Einstein's paper "On the Quantum Theory of Radiation" outlined the principles of the stimulated emission of photons. This led to the invention of the Laser (light amplification by the stimulated emission of radiation) and the optical amplifier which ushered in the Information Age. It is optical amplification that allows fiber optic networks to transmit the massive capacity of the Internet. Based on wireless transmission of electromagnetic radiation and global networks of cellular operation, the mobile phone became a primary means to access the internet. Developments in political science and economics In political science during the 20th century, the study of ideology, behaviouralism and international relations led to a multitude of 'pol-sci' subdisciplines including rational choice theory, voting theory, game theory (also used in economics), psephology, political geography/geopolitics, political anthropology/political psychology/political sociology, political economy, policy analysis, public administration, comparative political analysis and peace studies/conflict analysis. In economics, John Maynard Keynes prompted a division between microeconomics and macroeconomics in the 1920s. Under Keynesian economics macroeconomic trends can overwhelm economic choices made by individuals. Governments should promote aggregate demand for goods as a means to encourage economic expansion. Following World War II, Milton Friedman created the concept of monetarism. Monetarism focuses on using the supply and demand of money as a method for controlling economic activity. In the 1970s, monetarism has adapted into supply-side economics which advocates reducing taxes as a means to increase the amount of money available for economic expansion. Other modern schools of economic thought are New Classical economics and New Keynesian economics. New Classical economics was developed in the 1970s, emphasizing solid microeconomics as the basis for macroeconomic growth. New Keynesian economics was created partially in response to New Classical economics, and deals with how inefficiencies in the market create a need for control by a central bank or government. Developments in psychology, sociology, and anthropology Psychology in the 20th century saw a rejection of Freud's theories as being too unscientific, and a reaction against Edward Titchener's atomistic approach of the mind. This led to the formulation of behaviorism by John B. Watson, which was popularized by B.F. Skinner. Behaviorism proposed epistemologically limiting psychological study to overt behavior, since that could be reliably measured. Scientific knowledge of the "mind" was considered too metaphysical, hence impossible to achieve. The final decades of the 20th century have seen the rise of cognitive science, which considers the mind as once again a subject for investigation, using the tools of psychology, linguistics, computer science, philosophy, and neurobiology. New methods of visualizing the activity of the brain, such as PET scans and CAT scans, began to exert their influence as well, leading some researchers to investigate the mind by investigating the brain, rather than cognition. These new forms of investigation assume that a wide understanding of the human mind is possible, and that such an understanding may be applied to other research domains, such as artificial intelligence. Evolutionary theory was applied to behavior and introduced to anthropology and psychology through the works of cultural anthropologist Napoleon Chagnon; physical anthropology would eventually become evolutionary anthropology, incorporating elements of evolutionary biology with cultural anthropology. American sociology in the 1940s and 1950s was dominated largely by Talcott Parsons, who argued that aspects of society that promoted structural integration were therefore "functional". This structural functionalism approach was questioned in the 1960s, when sociologists came to see this approach as merely a justification for inequalities present in the status quo. In reaction, conflict theory was developed, which was based in part on the philosophies of Karl Marx. Conflict theorists saw society as an arena in which different groups compete for control over resources. Symbolic interactionism also came to be regarded as central to sociological thinking. Erving Goffman saw social interactions as a stage performance, with individuals preparing "backstage" and attempting to control their audience through impression management. While these theories are currently prominent in sociological thought, other approaches exist, including feminist theory, post-structuralism, rational choice theory, and postmodernism. In the mid-20th century, much of the methodologies of earlier anthropological and ethnographical study were reevaluated with an eye towards research ethics, while at the same time the scope of investigation has broadened far beyond the traditional study of "primitive cultures". 21st century Higgs boson On July 4, 2012, physicists working at CERN's Large Hadron Collider announced that they had discovered a new subatomic particle greatly resembling the Higgs boson, a potential key to an understanding of why elementary particles have mass and indeed to the existence of diversity and life in the universe. For now, some physicists are calling it a "Higgslike" particle. Peter Higgs was one of six physicists, working in three independent groups, who, in 1964, invented the notion of the Higgs field ("cosmic molasses"). The others were: Tom Kibble of Imperial College, London; Carl Hagen of the University of Rochester; Gerald Guralnik of Brown University; François Englert and Robert Brout, both of Université libre de Bruxelles. See also History History of mathematics History of measurement History of philosophy History of astronomy History of biology History of chemistry History of Earth science History of physics History of the social sciences History of science and technology 2020s in science and technology Science and technology in Canada History of science and technology in China Science and technology in India Timeline of science and technology in the Islamic world History of technology History and Philosophy of Science Philosophy of science Imre Lakatos Naïve empiricism History of scholarship Science studies History of science policy List of experiments List of multiple discoveries List of Nobel laureates List of people considered father or mother of a scientific field List of scientists List of years in science Multiple discovery Philosophy of history Science Fields of science Cognitive sciences Natural sciences Social sciences Science tourism Theories and sociology of the history of science Timelines of science Timeline of scientific discoveries Timeline of scientific experiments Timeline of scientific thought Timeline of the history of the scientific method Women in science References Sources Further reading Agar, Jon (2012) Science in the Twentieth Century and Beyond, Polity Press. . Agassi, Joseph (2007) Science and Its History: A Reassessment of the Historiography of Science (Boston Studies in the Philosophy of Science, 253) Springer. . Bowler, Peter J. (1993) The Norton History of the Environmental Sciences. Brock, W.H. (1993) The Norton History of Chemistry. Bronowski, J. (1951) The Common Sense of Science Heinemann. .) (Includes a description of the history of science in England.) Byers, Nina and Gary Williams, ed. (2006) Out of the Shadows: Contributions of Twentieth-Century Women to Physics, Cambridge University Press Herzenberg, Caroline L. (1986). Women Scientists from Antiquity to the Present Locust Hill Press Kumar, Deepak (2006). Science and the Raj: A Study of British India, 2nd edition. Oxford University Press. Lakatos, Imre (1978). History of Science and its Rational Reconstructions published in The Methodology of Scientific Research Programmes: Philosophical Papers Volume 1. Cambridge University Press Levere, Trevor Harvey. (2001) Transforming Matter: A History of Chemistry from Alchemy to the Buckyball Lipphardt, Veronika/Ludwig, Daniel, Knowledge Transfer and Science Transfer, EGO – European History Online, Mainz: Institute of European History, 2011, retrieved: March 8, 2020 (pdf). Margolis, Howard (2002). It Started with Copernicus. McGraw-Hill. Mayr, Ernst. (1985). The Growth of Biological Thought: Diversity, Evolution, and Inheritance. North, John. (1995). The Norton History of Astronomy and Cosmology. Nye, Mary Jo, ed. (2002). The Cambridge History of Science, Volume 5: The Modern Physical and Mathematical Sciences Park, Katharine, and Lorraine Daston, eds. (2006) The Cambridge History of Science, Volume 3: Early Modern Science Porter, Roy, ed. (2003). The Cambridge History of Science, Volume 4: The Eighteenth Century Rousseau, George and Roy Porter, eds. 1980). The Ferment of Knowledge: Studies in the Historiography of Science Cambridge University Press. Slotten, Hugh Richard, ed. (2014) The Oxford Encyclopedia of the History of American Science, Medicine, and Technology. External links 'What is the History of Science', British Academy British Society for the History of Science The CNRS History of Science and Technology Research Center in Paris (France) Henry Smith Williams, History of Science, Vols 1–4, online text Digital Archives of the National Institute of Standards and Technology (NIST) Digital facsimiles of books from the History of Science Collection, Linda Hall Library Digital Collections Division of History of Science and Technology of the International Union of History and Philosophy of Science Giants of Science (website of the Institute of National Remembrance) History of Science Digital Collection: Utah State University – Contains primary sources by such major figures in the history of scientific inquiry as Otto Brunfels, Charles Darwin, Erasmus Darwin, Carolus Linnaeus Antony van Leeuwenhoek, Jan Swammerdam, James Sowerby, Andreas Vesalius, and others. History of Science Society ("HSS") Inter-Divisional Teaching Commission (IDTC) of the International Union for the History and Philosophy of Science (IUHPS) International Academy of the History of Science International History, Philosophy and Science Teaching Group IsisCB Explore: History of Science Index An open access discovery tool Museo Galileo – Institute and Museum of the History of Science in Florence, Italy National Center for Atmospheric Research (NCAR) Archives The official site of the Nobel Foundation. Features biographies and info on Nobel laureates The Royal Society, trailblazing science from 1650 to date The Vega Science Trust Free to view videos of scientists including Feynman, Perutz, Rotblat, Born and many Nobel Laureates. Science studies
14617
https://en.wikipedia.org/wiki/Intel
Intel
Intel Corporation (commonly known as Intel) is an American multinational corporation and technology company headquartered in Santa Clara, California. It is one of the world's largest semiconductor chip manufacturer by revenue, and is one of the developers of the x86 series of instruction sets found in most personal computers (PCs). Incorporated in Delaware, Intel ranked No. 45 in the 2020 Fortune 500 list of the largest United States corporations by total revenue for nearly a decade, from 2007 to 2016 fiscal years. Intel supplies microprocessors for computer system manufacturers such as Acer, Lenovo, HP, and Dell. Intel also manufactures motherboard chipsets, network interface controllers and integrated circuits, flash memory, graphics chips, embedded processors and other devices related to communications and computing. Intel (integrated and electronics) was founded on July 18, 1968, by semiconductor pioneers Gordon Moore (of Moore's law) and Robert Noyce (1927–1990), and is associated with the executive leadership and vision of Andrew Grove. Intel was a key component of the rise of Silicon Valley as a high-tech center. Noyce was a key inventor of the integrated circuit (microchip). Intel was an early developer of SRAM and DRAM memory chips, which represented the majority of its business until 1981. Although Intel created the world's first commercial microprocessor chip in 1971, it was not until the success of the personal computer (PC) that this became its primary business. During the 1990s, Intel invested heavily in new microprocessor designs fostering the rapid growth of the computer industry. During this period, Intel became the dominant supplier of PC microprocessors and was known for aggressive and anti-competitive tactics in defense of its market position, particularly against AMD, as well as a struggle with Microsoft for control over the direction of the PC industry. The Open Source Technology Center at Intel hosts PowerTOP and LatencyTOP, and supports other open-source projects such as Wayland, Mesa, Threading Building Blocks (TBB), and Xen. Current operations Operating segments Client Computing Group 51.8% of 2020 revenues produces PC processors and related components. Data Center Group 33.7% of 2020 revenues produces hardware components used in server, network, and storage platforms. Non-Volatile Memory Solutions Group 6.9% of 2020 revenues produces components for solid-state drives: NAND flash memory and 3D XPoint (Optane). Internet of Things Group 5.2% of 2020 revenues offers platforms designed for retail, transportation, industrial, buildings and home use. Programmable Solutions Group 2.4% of 2020 revenues manufactures programmable semiconductors (primarily FPGAs). Customers In 2020, Dell accounted for about 17% of Intel's total revenues, Lenovo accounted for 12% of total revenues, and HP Inc. accounted for 10% of total revenues. As of August 2021, the US Department of Defense is another large customer for Intel. Market share According to IDC, while Intel enjoyed the biggest market share in both the overall worldwide PC microprocessor market (73.3%) and the mobile PC microprocessor (80.4%) in the second quarter of 2011, the numbers decreased by 1.5% and 1.9% compared to the first quarter of 2011. Intel's market share decreased significantly in the enthusiast market as of 2019, and they have faced delays for their 10 nm products. According to former Intel CEO Bob Swan, the delay was caused by the company's overly aggressive strategy for moving to its next node. Historical market share In the 1980s, Intel was among the world's top ten sellers of semiconductors (10th in 1987). It was part of the "Win-Tel" personal computer domination in the 1990s and early 2000s. In 1992, Intel became the biggest chip maker by revenue and held the position until 2018 when Samsung surpassed it, but Intel returned to its former position the year after. Other major semiconductor companies include TSMC, GlobalFoundries, Samsung, Texas Instruments, Toshiba, STMicroelectronics, United Microelectronics Corporation (UMC), Micron, SK Hynix, Kioxia, and SMIC. Major competitors Intel's competitors in PC chipsets included AMD, VIA Technologies, Silicon Integrated Systems, and Nvidia. Intel's competitors in networking include NXP Semiconductors, Infineon, Broadcom Limited, Marvell Technology Group and Applied Micro Circuits Corporation, and competitors in flash memory included Spansion, Samsung Electronics, Qimonda, Kioxia, STMicroelectronics, Micron, and SK Hynix. The only major competitor in the x86 processor market is AMD, with which Intel has had full cross-licensing agreements since 1976: each partner can use the other's patented technological innovations without charge after a certain time. However, the cross-licensing agreement is canceled in the event of an AMD bankruptcy or takeover. Some smaller competitors, such as VIA Technologies, produce low-power x86 processors for small factor computers and portable equipment. However, the advent of such mobile computing devices, in particular, smartphones, has in recent years led to a decline in PC sales. Since over 95% of the world's smartphones currently use processors cores designed by ARM Holdings, using the ARM instruction set, ARM has become a major competitor for Intel's processor market. ARM is also planning to make attempts at setting foot into the PC and server market, with Ampere and IBM each individually designing CPUs for servers and supercomputers. The only other major competitor in processor instruction sets is RISC-V, which is an open-source CPU instruction set. With the major phone and telecommunications manufacturer Huawei releasing chips based on the RISC instruction set due to US sanctions. Intel has been involved in several disputes regarding the violation of antitrust laws, which are noted below. Carbon footprint Intel reported Total CO2e emissions (Direct + Indirect) for the twelve months ending 31 December 2020 at 2,882 Kt (+94/+3.4% y-o-y). Intel plans to reduce carbon emissions 10% by 2030 from a 2020 base year. History Origins Intel was founded in Mountain View, California, on July 18, 1968 by Gordon E. Moore (known for "Moore's law"), a chemist, and Robert Noyce, a physicist and co-inventor of the integrated circuit. Arthur Rock (investor and venture capitalist) helped them find investors, while Max Palevsky was on the board from an early stage. Moore and Noyce had left Fairchild Semiconductor to found Intel. Rock was not an employee, but he was an investor and was chairman of the board. The total initial investment in Intel was $2.5 million in convertible debentures (equivalent to $ million in ) and $10,000 from Rock. Just 2 years later, Intel became a public company via an initial public offering (IPO), raising $6.8 million ($23.50 per share). Intel's third employee was Andy Grove, a chemical engineer, who later ran the company through much of the 1980s and the high-growth 1990s. In deciding on a name, Moore and Noyce quickly rejected "Moore Noyce", near homophone for "more noise" – an ill-suited name for an electronics company, since noise in electronics is usually undesirable and typically associated with bad interference. Instead, they founded the company as NM Electronics (or MN Electronics) on July 18, 1968, but by the end of the month had changed the name to Intel which stood for Integrated Electronics. Since "Intel" was already trademarked by the hotel chain Intelco, they had to buy the rights for the name. Early history At its founding, Intel was distinguished by its ability to make logic circuits using semiconductor devices. The founders' goal was the semiconductor memory market, widely predicted to replace magnetic-core memory. Its first product, a quick entry into the small, high-speed memory market in 1969, was the 3101 Schottky TTL bipolar 64-bit static random-access memory (SRAM), which was nearly twice as fast as earlier Schottky diode implementations by Fairchild and the Electrotechnical Laboratory in Tsukuba, Japan. In the same year, Intel also produced the 3301 Schottky bipolar 1024-bit read-only memory (ROM) and the first commercial metal–oxide–semiconductor field-effect transistor (MOSFET) silicon gate SRAM chip, the 256-bit 1101. While the 1101 was a significant advance, its complex static cell structure made it too slow and costly for mainframe memories. The three-transistor cell implemented in the first commercially available dynamic random-access memory (DRAM), the 1103 released in 1970, solved these issues. The 1103 was the bestselling semiconductor memory chip in the world by 1972, as it replaced core memory in many applications. Intel's business grew during the 1970s as it expanded and improved its manufacturing processes and produced a wider range of products, still dominated by various memory devices. Intel created the first commercially available microprocessor (Intel 4004) in 1971. The microprocessor represented a notable advance in the technology of integrated circuitry, as it miniaturized the central processing unit of a computer, which then made it possible for small machines to perform calculations that in the past only very large machines could do. Considerable technological innovation was needed before the microprocessor could actually become the basis of what was first known as a "mini computer" and then known as a "personal computer". Intel also created one of the first microcomputers in 1973. Intel opened its first international manufacturing facility in 1972, in Malaysia, which would host multiple Intel operations, before opening assembly facilities and semiconductor plants in Singapore and Jerusalem in the early 1980s, and manufacturing and development centers in China, India, and Costa Rica in the 1990s. By the early 1980s, its business was dominated by dynamic random-access memory (DRAM) chips. However, increased competition from Japanese semiconductor manufacturers had, by 1983, dramatically reduced the profitability of this market. The growing success of the IBM personal computer, based on an Intel microprocessor, was among factors that convinced Gordon Moore (CEO since 1975) to shift the company's focus to microprocessors and to change fundamental aspects of that business model. Moore's decision to sole-source Intel's 386 chip played into the company's continuing success. By the end of the 1980s, buoyed by its fortuitous position as microprocessor supplier to IBM and IBM's competitors within the rapidly growing personal computer market, Intel embarked on a 10-year period of unprecedented growth as the primary (and most profitable) hardware supplier to the PC industry, part of the winning 'Wintel' combination. Moore handed over his position as CEO to Andy Grove in 1987. By launching its Intel Inside marketing campaign in 1991, Intel was able to associate brand loyalty with consumer selection, so that by the end of the 1990s, its line of Pentium processors had become a household name. Challenges to dominance (2000s) After 2000, growth in demand for high-end microprocessors slowed. Competitors, most notably AMD (Intel's largest competitor in its primary x86 architecture market), garnered significant market share, initially in low-end and mid-range processors but ultimately across the product range, and Intel's dominant position in its core market was greatly reduced, mostly due to controversial NetBurst microarchitecture. In the early 2000s then-CEO, Craig Barrett attempted to diversify the company's business beyond semiconductors, but few of these activities were ultimately successful. Litigation Intel had also for a number of years been embroiled in litigation. US law did not initially recognize intellectual property rights related to microprocessor topology (circuit layouts), until the Semiconductor Chip Protection Act of 1984, a law sought by Intel and the Semiconductor Industry Association (SIA). During the late 1980s and 1990s (after this law was passed), Intel also sued companies that tried to develop competitor chips to the 80386 CPU. The lawsuits were noted to significantly burden the competition with legal bills, even if Intel lost the suits. Antitrust allegations had been simmering since the early 1990s and had been the cause of one lawsuit against Intel in 1991. In 2004 and 2005, AMD brought further claims against Intel related to unfair competition. Reorganization and success with Intel Core (2005–2015) In 2005, CEO Paul Otellini reorganized the company to refocus its core processor and chipset business on platforms (enterprise, digital home, digital health, and mobility). On June 6, 2005, Steve Jobs, then CEO of Apple, announced that Apple would be using Intel's x86 processors for its Macintosh computers, switching from the PowerPC architecture developed by the AIM alliance. This was seen as win for Intel, although an analyst called the move "risky" and "foolish", as Intel's current offerings at the time were considered to be behind those of AMD and IBM. In 2006, Intel unveiled its Core microarchitecture to widespread critical acclaim; the product range was perceived as an exceptional leap in processor performance that at a stroke regained much of its leadership of the field. In 2008, Intel had another "tick" when it introduced the Penryn microarchitecture, fabricated using the 45 nm process node. Later that year, Intel released a processor with the Nehalem architecture to positive reception. On June 27, 2006, the sale of Intel's XScale assets was announced. Intel agreed to sell the XScale processor business to Marvell Technology Group for an estimated $600 million and the assumption of unspecified liabilities. The move was intended to permit Intel to focus its resources on its core x86 and server businesses, and the acquisition completed on November 9, 2006. In 2008, Intel spun off key assets of a solar startup business effort to form an independent company, SpectraWatt Inc. In 2011, SpectraWatt filed for bankruptcy. In February 2011, Intel began to build a new microprocessor manufacturing facility in Chandler, Arizona, completed in 2013 at a cost of $5 billion. The building is now the 10 nm-certified Fab 42 and is connected to the other Fabs (12, 22, 32) on Ocotillo Campus via an enclosed bridge known as the Link. The company produces three-quarters of its products in the United States, although three-quarters of its revenue come from overseas. The Alliance for Affordable Internet (A4AI) was launched in October 2013 and Intel is part of the coalition of public and private organizations that also includes Facebook, Google, and Microsoft. Led by Sir Tim Berners-Lee, the A4AI seeks to make Internet access more affordable so that access is broadened in the developing world, where only 31% of people are online. Google will help to decrease Internet access prices so that they fall below the UN Broadband Commission's worldwide target of 5% of monthly income. Attempts at entering the smartphone market In April 2011, Intel began a pilot project with ZTE Corporation to produce smartphones using the Intel Atom processor for China's domestic market. In December 2011, Intel announced that it reorganized several of its business units into a new mobile and communications group that would be responsible for the company's smartphone, tablet, and wireless efforts. Intel planned to introduce Medfield – a processor for tablets and smartphones – to the market in 2012, as an effort to compete with ARM. As a 32-nanometer processor, Medfield is designed to be energy-efficient, which is one of the core features in ARM's chips. At the Intel Developers Forum (IDF) 2011 in San Francisco, Intel's partnership with Google was announced. In January 2012, Google announced Android 2.3, supporting Intel's Atom microprocessor. In 2013, Intel's Kirk Skaugen said that Intel's exclusive focus on Microsoft platforms was a thing of the past and that they would now support all "tier-one operating systems" such as Linux, Android, iOS, and Chrome. In 2014, Intel cut thousands of employees in response to "evolving market trends", and offered to subsidize manufacturers for the extra costs involved in using Intel chips in their tablets. In April 2016, Intel cancelled the SoFIA platform and the Broxton Atom SoC for smartphones, effectively leaving the smartphone market. Intel custom foundry Finding itself with excess fab capacity after the failure of the Ultrabook to gain market traction and with PC sales declining, in 2013 Intel reached a foundry agreement to produce chips for Altera using a 14nm process. General Manager of Intel's custom foundry division Sunit Rikhi indicated that Intel would pursue further such deals in the future. This was after poor sales of Windows 8 hardware caused a major retrenchment for most of the major semiconductor manufacturers, except for Qualcomm, which continued to see healthy purchases from its largest customer, Apple. As of July 2013, five companies were using Intel's fabs via the Intel Custom Foundry division: Achronix, Tabula, Netronome, Microsemi, and Panasonic most are field-programmable gate array (FPGA) makers, but Netronome designs network processors. Only Achronix began shipping chips made by Intel using the 22nm Tri-Gate process. Several other customers also exist but were not announced at the time. The foundry business was closed in 2018 due to Intel's issues with its manufacturing. Security and manufacturing challenges (2016–2021) Intel continued its tick-tock model of a microarchitecture change followed by a die shrink until the 6th generation Core family based on the Skylake microarchitecture. This model was deprecated in 2016, with the release of the seventh generation Core family (codenamed Kaby Lake), ushering in the process–architecture–optimization model. As Intel struggled to shrink their process node from 14 nm to 10 nm, processor development slowed down and the company continued to use the Skylake microarchitecture until 2020, albeit with optimizations. 10 nm process node issues While Intel originally planned to introduce 10 nm products in 2016, it later became apparent that there were manufacturing issues with the node. The first microprocessor under that node, Cannon Lake (marketed as 8th generation Core), was released in small quantities in 2018. The company first delayed the mass production of their 10 nm products to 2017. They later delayed mass production to 2018, and then to 2019. Despite rumors of the process being cancelled, Intel finally introduced mass-produced 10 nm 10th generation Intel Core mobile processors (codenamed "Ice Lake") in September 2019. Intel later acknowledged that their strategy to shrink to 10 nm was too aggressive. While other foundries used up to four steps in 10 nm or 7 nm processes, the company's 10 nm process required up to five or six multi-pattern steps. In addition, Intel's 10 nm process is denser than its counterpart processes from other foundries. Since Intel's microarchitecture and process node development were coupled, processor development stagnated. Security flaws In early January 2018, it was reported that all Intel processors made since 1995, excluding Intel Itanium and pre-2013 Intel Atom processors, have been subject to two security flaws dubbed Meltdown and Spectre. It is believed that "hundreds of millions" of systems could be affected by these flaws. More security flaws were disclosed on May 3, 2018, on August 14, 2018, on January 18, 2019, and on March 5, 2020. On March 15, 2018, Intel reported that it will redesign its CPUs to protect against the Spectre security vulnerability, the redesigned processors were sold later in 2018. Existing chips vulnerable to Meltdown and Spectre can be fixed with a software patch at a cost to performance. Renewed competition and other developments (2018–present) Due to Intel's issues with its 10 nm process node and the company's slow processor development, the company now found itself in a market with intense competition. The company's main competitor, AMD, introduced the Zen microarchitecture and a new chiplet based design to critical acclaim. Since its introduction, AMD, once unable to compete with Intel in the high-end CPU market, has undergone a resurgence, and Intel's dominance and market share have considerably decreased. In addition, Apple is switching from the x86 architecture and Intel processors to their own Apple silicon for their Macintosh computers from 2020 onwards. The transition is expected to affect Intel minimally; however, it might prompt other PC manufacturers to reevaluate their reliance on Intel and the x86 architecture. 'IDM 2.0' strategy On March 23, 2021, CEO Pat Gelsinger laid out new plans for the company. These include a new strategy, called IDM 2.0, that includes investments in manufacturing facilities, use of both internal and external foundries, and a new foundry business called Intel Foundry Services (IFS), a standalone business unit. Unlike Intel Custom Foundry, IFS will offer a combination of packaging and process technology, and Intel's IP portfolio including x86 cores. Other plans for the company include a partnership with IBM and a new event for developers and engineers, called "Intel ON". Gelsinger also confirmed that Intel's 7 nm process is on track, and that the first products using their 7 nm process (also known as Intel 4) are Ponte Vecchio and Meteor Lake. In January 2022, Intel reportedly selected New Albany, Ohio, near Columbus, Ohio, as the site for a major new manufacturing facility. The facility will cost at least $20 billion. The company expects the facility to begin producing chips by 2025. The same year Intel also choose Magdeburg, Germany, as a site for two new chip mega factories for €17 billion (topping Tesla's investment in Brandenburg). Groundbreaking is planned for 2023, while production start is planned for 2027. Including subcontractors this would create 10.000 new jobs. In August 2022, Intel signed a $30billion partnership with Brookfield Asset Management to fund its recent factory expansions. As part of the deal, Intel would have a controlling stake by funding 51% of the cost of building new chip-making facilities in Chandler, with Brookfield owning the remaining 49% stake, allowing the companies to split the revenue from those facilities. On January 31, 2023, as part of $3 billion in cost reductions, Intel announced pay cuts affecting employees above midlevel, ranging from 5% upwards. It also suspended bonuses and merit pay increases, while reducing retirement plan matching. These cost reductions followed layoffs announced in the fall of 2022. Product and market history SRAMs, DRAMs, and the microprocessor Intel's first products were shift register memory and random-access memory integrated circuits, and Intel grew to be a leader in the fiercely competitive DRAM, SRAM, and ROM markets throughout the 1970s. Concurrently, Intel engineers Marcian Hoff, Federico Faggin, Stanley Mazor, and Masatoshi Shima invented Intel's first microprocessor. Originally developed for the Japanese company Busicom to replace a number of ASICs in a calculator already produced by Busicom, the Intel 4004 was introduced to the mass market on November 15, 1971, though the microprocessor did not become the core of Intel's business until the mid-1980s. (Note: Intel is usually given credit with Texas Instruments for the almost-simultaneous invention of the microprocessor) In 1983, at the dawn of the personal computer era, Intel's profits came under increased pressure from Japanese memory-chip manufacturers, and then-president Andy Grove focused the company on microprocessors. Grove described this transition in the book Only the Paranoid Survive. A key element of his plan was the notion, then considered radical, of becoming the single source for successors to the popular 8086 microprocessor. Until then, the manufacture of complex integrated circuits was not reliable enough for customers to depend on a single supplier, but Grove began producing processors in three geographically distinct factories, and ceased licensing the chip designs to competitors such as AMD. When the PC industry boomed in the late 1980s and 1990s, Intel was one of the primary beneficiaries. Early x86 processors and the IBM PC Despite the ultimate importance of the microprocessor, the 4004 and its successors the 8008 and the 8080 were never major revenue contributors at Intel. As the next processor, the 8086 (and its variant the 8088) was completed in 1978, Intel embarked on a major marketing and sales campaign for that chip nicknamed "Operation Crush", and intended to win as many customers for the processor as possible. One design win was the newly created IBM PC division, though the importance of this was not fully realized at the time. IBM introduced its personal computer in 1981, and it was rapidly successful. In 1982, Intel created the 80286 microprocessor, which, two years later, was used in the IBM PC/AT. Compaq, the first IBM PC "clone" manufacturer, produced a desktop system based on the faster 80286 processor in 1985 and in 1986 quickly followed with the first 80386-based system, beating IBM and establishing a competitive market for PC-compatible systems and setting up Intel as a key component supplier. In 1975, the company had started a project to develop a highly advanced 32-bit microprocessor, finally released in 1981 as the Intel iAPX 432. The project was too ambitious and the processor was never able to meet its performance objectives, and it failed in the marketplace. Intel extended the x86 architecture to 32 bits instead. 386 microprocessor During this period Andrew Grove dramatically redirected the company, closing much of its DRAM business and directing resources to the microprocessor business. Of perhaps greater importance was his decision to "single-source" the 386 microprocessor. Prior to this, microprocessor manufacturing was in its infancy, and manufacturing problems frequently reduced or stopped production, interrupting supplies to customers. To mitigate this risk, these customers typically insisted that multiple manufacturers produce chips they could use to ensure a consistent supply. The 8080 and 8086-series microprocessors were produced by several companies, notably AMD, with which Intel had a technology-sharing contract. Grove made the decision not to license the 386 design to other manufacturers, instead, producing it in three geographically distinct factories: Santa Clara, California; Hillsboro, Oregon; and Chandler, a suburb of Phoenix, Arizona. He convinced customers that this would ensure consistent delivery. In doing this, Intel breached its contract with AMD, which sued and was paid millions of dollars in damages but could not manufacture new Intel CPU designs any longer. (Instead, AMD started to develop and manufacture its own competing x86 designs.) As the success of Compaq's Deskpro 386 established the 386 as the dominant CPU choice, Intel achieved a position of near-exclusive dominance as its supplier. Profits from this funded rapid development of both higher-performance chip designs and higher-performance manufacturing capabilities, propelling Intel to a position of unquestioned leadership by the early 1990s. 486, Pentium, and Itanium Intel introduced the 486 microprocessor in 1989, and in 1990 established a second design team, designing the processors code-named "P5" and "P6" in parallel and committing to a major new processor every two years, versus the four or more years such designs had previously taken. Engineers Vinod Dham and Rajeev Chandrasekhar (Member of Parliament, India) were key figures on the core team that invented the 486 chip and later, Intel's signature Pentium chip. The P5 project was earlier known as "Operation Bicycle", referring to the cycles of the processor through two parallel execution pipelines. The P5 was introduced in 1993 as the Intel Pentium, substituting a registered trademark name for the former part number (numbers, such as 486, cannot be legally registered as trademarks in the United States). The P6 followed in 1995 as the Pentium Pro and improved into the Pentium II in 1997. New architectures were developed alternately in Santa Clara, California and Hillsboro, Oregon. The Santa Clara design team embarked in 1993 on a successor to the x86 architecture, codenamed "P7". The first attempt was dropped a year later but quickly revived in a cooperative program with Hewlett-Packard engineers, though Intel soon took over primary design responsibility. The resulting implementation of the IA-64 64-bit architecture was the Itanium, finally introduced in June 2001. The Itanium's performance running legacy x86 code did not meet expectations, and it failed to compete effectively with x86-64, which was AMD's 64-bit extension of the 32-bit x86 architecture (Intel uses the name Intel 64, previously EM64T). In 2017, Intel announced that the Itanium 9700 series (Kittson) would be the last Itanium chips produced. The Hillsboro team designed the Willamette processors (initially code-named P68), which were marketed as the Pentium 4. During this period, Intel undertook two major supporting advertising campaigns. The first campaign, the 1991 "Intel Inside" marketing and branding campaign, is widely known and has become synonymous with Intel itself. The idea of "ingredient branding" was new at the time, with only NutraSweet and a few others making attempts to do so. This campaign established Intel, which had been a component supplier little-known outside the PC industry, as a household name. The second campaign, Intel's Systems Group, which began in the early 1990s, showcased manufacturing of PC motherboards, the main board component of a personal computer, and the one into which the processor (CPU) and memory (RAM) chips are plugged. The Systems Group campaign was lesser known than the Intel Inside campaign. Shortly after, Intel began manufacturing fully configured "white box" systems for the dozens of PC clone companies that rapidly sprang up. At its peak in the mid-1990s, Intel manufactured over 15% of all PCs, making it the third-largest supplier at the time. During the 1990s, Intel Architecture Labs (IAL) was responsible for many of the hardware innovations for the PC, including the PCI Bus, the PCI Express (PCIe) bus, and Universal Serial Bus (USB). IAL's software efforts met with a more mixed fate; its video and graphics software was important in the development of software digital video, but later its efforts were largely overshadowed by competition from Microsoft. The competition between Intel and Microsoft was revealed in testimony by then IAL Vice-president Steven McGeady at the Microsoft antitrust trial (United States v. Microsoft Corp.). Pentium flaw In June 1994, Intel engineers discovered a flaw in the floating-point math subsection of the P5 Pentium microprocessor. Under certain data-dependent conditions, the low-order bits of the result of a floating-point division would be incorrect. The error could compound in subsequent calculations. Intel corrected the error in a future chip revision, and under public pressure it issued a total recall and replaced the defective Pentium CPUs (which were limited to some 60, 66, 75, 90, and 100 MHz models) on customer request. The bug was discovered independently in October 1994 by Thomas Nicely, Professor of Mathematics at Lynchburg College. He contacted Intel but received no response. On October 30, he posted a message about his finding on the Internet. Word of the bug spread quickly and reached the industry press. The bug was easy to replicate; a user could enter specific numbers into the calculator on the operating system. Consequently, many users did not accept Intel's statements that the error was minor and "not even an erratum". During Thanksgiving, in 1994, The New York Times ran a piece by journalist John Markoff spotlighting the error. Intel changed its position and offered to replace every chip, quickly putting in place a large end-user support organization. This resulted in a $475 million charge against Intel's 1994 revenue. Dr. Nicely later learned that Intel had discovered the FDIV bug in its own testing a few months before him (but had decided not to inform customers). The "Pentium flaw" incident, Intel's response to it, and the surrounding media coverage propelled Intel from being a technology supplier generally unknown to most computer users to a household name. Dovetailing with an uptick in the "Intel Inside" campaign, the episode is considered to have been a positive event for Intel, changing some of its business practices to be more end-user focused and generating substantial public awareness, while avoiding a lasting negative impression. Intel Core The Intel Core line originated from the original Core brand, with the release of the 32-bit Yonah CPU, Intel's first dual-core mobile (low-power) processor. Derived from the Pentium M, the processor family used an enhanced version of the P6 microarchitecture. Its successor, the Core 2 family, was released on July 27, 2006. This was based on the Intel Core microarchitecture, and was a 64-bit design. Instead of focusing on higher clock rates, the Core microarchitecture emphasized power efficiency and a return to lower clock speeds. It also provided more efficient decoding stages, execution units, caches, and buses, reducing the power consumption of Core 2-branded CPUs while increasing their processing capacity. In November 2008, Intel released the first generation Core processors based on the Nehalem microarchitecture. Intel also introduced a new naming scheme, with the three variants now named Core i3, i5, and i7 (as well as i9 from 7th generation onwards). Unlike the previous naming scheme, these names no longer correspond to specific technical features. It was succeeded by the Westmere microarchitecture in 2010, with a die shrink to 32 nm and included Intel HD Graphics. In 2011, Intel released the Sandy Bridge-based 2nd generation Core processor family. This generation featured an 11% performance increase over Nehalem. It was succeeded by Ivy Bridge-based 3rd generation Core, introduced at the 2012 Intel Developer Forum. Ivy Bridge featured a die shrink to 22 nm, and supported both DDR3 memory and DDR3L chips. Intel continued its tick-tock model of a microarchitecture change followed by a die shrink until the 6th generation Core family based on the Skylake microarchitecture. This model was deprecated in 2016, with the release of the seventh generation Core family based on Kaby Lake, ushering in the process–architecture–optimization model. From 2016 until 2021, Intel later released more optimizations on the Skylake microarchitecture with Kaby Lake R, Amber Lake, Whiskey Lake, Coffee Lake, Coffee Lake R, and Comet Lake. Intel struggled to shrink their process node from 14 nm to 10 nm, with the first microarchitecture under that node, Cannon Lake (marketed as 8th generation Core), only being released in small quantities in 2018. In 2019, Intel released the 10th generation of Core processors, codenamed "Amber Lake", "Comet Lake", and "Ice Lake". Ice Lake, based on the Sunny Cove microarchitecture, was produced on the 10 nm process and was limited to low-power mobile processors. Both Amber Lake and Comet Lake were based on a refined 14 nm node, with the latter being used for desktop and high performance mobile products and the former used for low-power mobile products. In September 2020, 11th generation Core mobile processors, codenamed Tiger Lake, were launched. Tiger Lake is based on the Willow Cove microarchitecture and a refined 10 nm node. Intel later released 11th generation Core desktop processors (codenamed "Rocket Lake"), fabricated using Intel's 14 nm process and based on the Cypress Cove microarchitecture, on March 30, 2021. It replaced Comet Lake desktop processors. All 11th generation Core processors feature new integrated graphics based on the Intel Xe microarchitecture. Both desktop and mobile products were unified under a single process node with the release of 12th generation Intel Core processors (codenamed "Alder Lake") in late 2021. This generation will be fabricated using Intel's 7 nm process, called Intel 7, for both desktop and mobile processors, and is based on a hybrid architecture utilizing high-performance Golden Cove cores and high-efficiency Gracemont (Atom) cores. Meltdown, Spectre, and other security vulnerabilities In early January 2018, it was reported that all Intel processors made since 1995 (besides Intel Itanium and pre-2013 Intel Atom) have been subject to two security flaws dubbed Meltdown and Spectre. The impact on performance resulting from software patches is "workload-dependent". Several procedures to help protect home computers and related devices from the Spectre and Meltdown security vulnerabilities have been published. Spectre patches have been reported to significantly slow down performance, especially on older computers; on the newer 8th generation Core platforms, benchmark performance drops of 2–14 percent have been measured. Meltdown patches may also produce performance loss. It is believed that "hundreds of millions" of systems could be affected by these flaws. On March 15, 2018, Intel reported that it will redesign its CPUs (performance losses to be determined) to protect against the Spectre security vulnerability, and expects to release the newly redesigned processors later in 2018. On May 3, 2018, eight additional Spectre-class flaws were reported. Intel reported that they are preparing new patches to mitigate these flaws. On August 14, 2018, Intel disclosed three additional chip flaws referred to as L1 Terminal Fault (L1TF). They reported that previously released microcode updates, along with new, pre-release microcode updates can be used to mitigate these flaws. On January 18, 2019, Intel disclosed three new vulnerabilities affecting all Intel CPUs, named "Fallout", "RIDL", and "ZombieLoad", allowing a program to read information recently written, read data in the line-fill buffers and load ports, and leak information from other processes and virtual machines. Coffeelake-series CPUs are even more vulnerable, due to hardware mitigations for Spectre. On March 5, 2020, computer security experts reported another Intel chip security flaw, besides the Meltdown and Spectre flaws, with the systematic name (or, "Intel CSME Bug"). This newly found flaw is not fixable with a firmware update, and affects nearly "all Intel chips released in the past five years". Use of Intel products by Apple Inc. (2005–2019) On June 6, 2005, Steve Jobs, then CEO of Apple, announced that Apple would be transitioning the Macintosh from its long favored PowerPC architecture to the Intel x86 architecture because the future PowerPC road map was unable to satisfy Apple's needs. This was seen as a win for Intel, although an analyst called the move "risky" and "foolish", as Intel's current offerings at the time were considered to be behind those of AMD and IBM. The first Mac computers containing Intel CPUs were announced on January 10, 2006, and Apple had its entire line of consumer Macs running on Intel processors by early August 2006. The Apple Xserve server was updated to Intel Xeon processors from November 2006 and was offered in a configuration similar to Apple's Mac Pro. Despite Apple's use of Intel products, relations between the two companies were strained at times. Rumors of Apple switching from Intel processors to their own designs began circulating as early as 2011. On June 22, 2020, during Apple's annual WWDC, Tim Cook, Apple's CEO, announced that it would be transitioning the company’s entire Mac line from Intel CPUs to custom Apple-designed processors based on the ARM architecture over the course of the next two years. In the short term, this transition was estimated to have minimal effects on Intel, as Apple only accounted for 2% to 4% of its revenue. However, Apple's shift to its own chips might prompt other PC manufacturers to reassess their reliance on Intel and the x86 architecture. By November 2020, Apple unveiled the M1, its processor custom-designed for the Mac. Solid-state drives (SSD) In 2008, Intel began shipping mainstream solid-state drives (SSDs) with up to 160 GB storage capacities. As with their CPUs, Intel develops SSD chips using ever-smaller nanometer processes. These SSDs make use of industry standards such as NAND flash, mSATA, PCIe, and NVMe. In 2017, Intel introduced SSDs based on 3D XPoint technology under the Optane brand name. In 2021, SK Hynix acquired most of Intel's NAND memory business for $7 billion, with a remaining transaction worth $2 billion expected in 2025. Intel also discontinued its consumer Optane products in 2021. In July 2022, Intel disclosed in its Q2 earnings report that it would cease future product development within its Optane business. Supercomputers The Intel Scientific Computers division was founded in 1984 by Justin Rattner, to design and produce parallel computers based on Intel microprocessors connected in hypercube internetwork topology. In 1992, the name was changed to the Intel Supercomputing Systems Division, and development of the iWarp architecture was also subsumed. The division designed several supercomputer systems, including the Intel iPSC/1, iPSC/2, iPSC/860, Paragon and ASCI Red. In November 2014, Intel stated that it was planning to use optical fibers to improve networking within supercomputers. Fog computing On November 19, 2015, Intel, alongside ARM Holdings, Dell, Cisco Systems, Microsoft, and Princeton University, founded the OpenFog Consortium, to promote interests and development in fog computing. Intel's Chief Strategist for the IoT Strategy and Technology Office, Jeff Fedders, became the consortium's first president. Self-driving cars Intel is one of the biggest stakeholders in the self-driving car industry, having joined the race in mid 2017 after joining forces with Mobileye. The company is also one of the first in the sector to research consumer acceptance, after an AAA report quoted a 78% nonacceptance rate of the technology in the US. Safety levels of autonomous driving technology, the thought of abandoning control to a machine, and psychological comfort of passengers in such situations were the major discussion topics initially. The commuters also stated that they did not want to see everything the car was doing. This was primarily a referral to the auto-steering wheel with no one sitting in the driving seat. Intel also learned that voice control regulator is vital, and the interface between the humans and machine eases the discomfort condition, and brings some sense of control back. It is important to mention that Intel included only 10 people in this study, which makes the study less credible. In a video posted on YouTube, Intel accepted this fact and called for further testing. Programmable devices Intel has sold Stratix, Arria, and Cyclone FPGAs since acquiring Altera in 2015. In 2019, Intel released Agilex FPGAs: chips aimed at data centers, 5G applications, and other uses. Competition, antitrust and espionage By the end of the 1990s, microprocessor performance had outstripped software demand for that CPU power. Aside from high-end server systems and software, whose demand dropped with the end of the "dot-com bubble", consumer systems ran effectively on increasingly low-cost systems after 2000. Intel's strategy of producing ever-more-powerful processors and obsoleting their predecessors stumbled, leaving an opportunity for rapid gains by competitors, notably AMD. This, in turn, lowered the profitability of the processor line and ended an era of unprecedented dominance of the PC hardware by Intel. Intel's dominance in the x86 microprocessor market led to numerous charges of antitrust violations over the years, including FTC investigations in both the late 1980s and in 1999, and civil actions such as the 1997 suit by Digital Equipment Corporation (DEC) and a patent suit by Intergraph. Intel's market dominance (at one time it controlled over 85% of the market for 32-bit x86 microprocessors) combined with Intel's own hardball legal tactics (such as its infamous 338 patent suit versus PC manufacturers) made it an attractive target for litigation, culminating in Intel agreeing to pay AMD $1.25B and grant them a perpetual patent cross-license in 2009 as well as several anti-trust judgements in Europe, Korea, and Japan. A case of industrial espionage arose in 1995 that involved both Intel and AMD. Bill Gaede, an Argentine formerly employed both at AMD and at Intel's Arizona plant, was arrested for attempting in 1993 to sell the i486 and P5 Pentium designs to AMD and to certain foreign powers. Gaede videotaped data from his computer screen at Intel and mailed it to AMD, which immediately alerted Intel and authorities, resulting in Gaede's arrest. Gaede was convicted and sentenced to 33 months in prison in June 1996. Corporate affairs Leadership and corporate structure Robert Noyce was Intel's CEO at its founding in 1968, followed by co-founder Gordon Moore in 1975. Andy Grove became the company's president in 1979 and added the CEO title in 1987 when Moore became chairman. In 1998, Grove succeeded Moore as chairman, and Craig Barrett, already company president, took over. On May 18, 2005, Barrett handed the reins of the company over to Paul Otellini, who had been the company president and COO and who was responsible for Intel's design win in the original IBM PC. The board of directors elected Otellini as president and CEO, and Barrett replaced Grove as Chairman of the Board. Grove stepped down as chairman but is retained as a special adviser. In May 2009, Barrett stepped down as chairman of the board and was succeeded by Jane Shaw. In May 2012, Intel vice chairman Andy Bryant, who had held the posts of CFO (1994) and Chief Administrative Officer (2007) at Intel, succeeded Shaw as executive chairman. In November 2012, president and CEO Paul Otellini announced that he would step down in May 2013 at the age of 62, three years before the company's mandatory retirement age. During a six-month transition period, Intel's board of directors commenced a search process for the next CEO, in which it considered both internal managers and external candidates such as Sanjay Jha and Patrick Gelsinger. Financial results revealed that, under Otellini, Intel's revenue increased by 55.8 percent (US$34.2 to 53.3 billion), while its net income increased by 46.7% (US$7.5 billion to 11 billion), proving that his illegal business practices were more profitable than the fines levied against the company as punishment for employing them. On May 2, 2013, Executive Vice President and COO Brian Krzanich was elected as Intel's sixth CEO, a selection that became effective on May 16, 2013, at the company's annual meeting. Reportedly, the board concluded that an insider could proceed with the role and exert an impact more quickly, without the need to learn Intel's processes, and Krzanich was selected on such a basis. Intel's software head Renée James was selected as president of the company, a role that is second to the CEO position. As of May 2013, Intel's board of directors consists of Andy Bryant, John Donahoe, Frank Yeary, Ambassador Charlene Barshefsky, Susan Decker, Reed Hundt, Paul Otellini, James Plummer, David Pottruck, and David Yoffie and Creative director will.i.am. The board was described by former Financial Times journalist Tom Foremski as "an exemplary example of corporate governance of the highest order" and received a rating of ten from GovernanceMetrics International, a form of recognition that has only been awarded to twenty-one other corporate boards worldwide. On June 21, 2018, Intel announced the resignation of Brian Krzanich as CEO, with the exposure of a relationship he had with an employee. Bob Swan was named interim CEO, as the Board began a search for a permanent CEO. On January 31, 2019, Swan transitioned from his role as CFO and interim CEO and was named by the Board as the seventh CEO to lead the company. On January 13, 2021, Intel announced that Swan would be replaced as CEO by Pat Gelsinger, effective February 15. Gelsinger is a former Intel chief technology officer who had previously been head of VMWare. Board of directors : Frank D. Yeary (chairman), managing member of Darwin Capital Pat Gelsinger, CEO of Intel James Goetz, managing director of Sequoia Capital Andrea Goldsmith, dean of engineering and applied science at Princeton University Alyssa Henry, Square, Inc. executive Omar Ishrak, chairman and former CEO of Medtronic Risa Lavizzo-Mourey, former president and CEO of the Robert Wood Johnson Foundation Tsu-Jae King Liu, professor at the UC Berkeley College of Engineering Barbara G. Novick, co-founder of BlackRock Gregory Smith, CFO of Boeing Dion Weisler, former president and CEO of HP Inc. Lip-Bu Tan, executive chairman of Cadence Design Systems Employment Intel has a mandatory retirement policy for its CEOs when they reach age 65. Andy Grove retired at 62, while both Robert Noyce and Gordon Moore retired at 58. Grove retired as chairman and as a member of the board of directors in 2005 at age 68. Intel's headquarters are located in Santa Clara, California, and the company has operations around the world. Its largest workforce concentration anywhere is in Washington County, Oregon (in the Portland metropolitan area's "Silicon Forest"), with 18,600 employees at several facilities. Outside the United States, the company has facilities in China, Costa Rica, Malaysia, Israel, Ireland, India, Russia, Argentina and Vietnam, in 63 countries and regions internationally. In March 2022, because of International sanctions during the Russo-Ukrainian War, Intel stopped supplying Russian market. In the U.S. Intel employs significant numbers of people in California, Colorado, Massachusetts, Arizona, New Mexico, Oregon, Texas, Washington and Utah. In Oregon, Intel is the state's largest private employer. The company is the largest industrial employer in New Mexico while in Arizona the company has 12,000 employees as of January 2020. Intel invests heavily in research in China and about 100 researchersor 10% of the total number of researchers from Intelare located in Beijing. In 2011, the Israeli government offered Intel $290 million to expand in the country. As a condition, Intel would employ 1,500 more workers in Kiryat Gat and between 600 and 1000 workers in the north. In January 2014, it was reported that Intel would cut about 5,000 jobs from its work force of 107,000. The announcement was made a day after it reported earnings that missed analyst targets. In March 2014, it was reported that Intel would embark upon a $6 billion plan to expand its activities in Israel. The plan calls for continued investment in existing and new Intel plants until 2030. , Intel employs 10,000 workers at four development centers and two production plants in Israel. Due to declining PC sales, in 2016 Intel cut 12,000 jobs. In 2021, Intel reversed course under new CEO Pat Gelsinger and started hiring thousands of engineers. Diversity Intel has a Diversity Initiative, including employee diversity groups as well as supplier diversity programs. Like many companies with employee diversity groups, they include groups based on race and nationality as well as sexual identity and religion. In 1994, Intel sanctioned one of the earliest corporate Gay, Lesbian, Bisexual, and Transgender employee groups, and supports a Muslim employees group, a Jewish employees group, and a Bible-based Christian group. Intel has received a 100% rating on numerous Corporate Equality Indices released by the Human Rights Campaign including the first one released in 2002. In addition, the company is frequently named one of the 100 Best Companies for Working Mothers by Working Mother magazine. In January 2015, Intel announced the investment of $300 million over the next five years to enhance gender and racial diversity in their own company as well as the technology industry as a whole. In February 2016, Intel released its Global Diversity & Inclusion 2015 Annual Report. The male-female mix of US employees was reported as 75.2% men and 24.8% women. For US employees in technical roles, the mix was reported as 79.8% male and 20.1% female. NPR reports that Intel is facing a retention problem (particularly for African Americans), not just a pipeline problem. Economic impact in Oregon in 2009 In 2011, ECONorthwest conducted an economic impact analysis of Intel's economic contribution to the state of Oregon. The report found that in 2009 "the total economic impacts attributed to Intel's operations, capital spending, contributions and taxes amounted to almost $14.6 billion in activity, including $4.3 billion in personal income and 59,990 jobs". Through multiplier effects, every 10 Intel jobs supported, on average, was found to create 31 jobs in other sectors of the economy. Intel Israel Intel has been operating in the State of Israel since Dov Frohman founded the Israeli branch of the company in 1974 in a small office in Haifa. Intel Israel currently has development centers in Haifa, Jerusalem and Petah Tikva, and has a manufacturing plant in the Kiryat Gat industrial park that develops and manufactures microprocessors and communications products. Intel employed about 10,000 employees in Israel in 2013. Maxine Fesberg has been the CEO of Intel Israel since 2007 and the Vice President of Intel Global. In December 2016, Fesberg announced her resignation, her position of chief executive officer (CEO) has been filled by Yaniv Gerti since January 2017. Key acquisitions and investments (2010–present) In 2010, Intel purchased McAfee, a manufacturer of computer security technology, for $7.68 billion. As a condition for regulatory approval of the transaction, Intel agreed to provide rival security firms with all necessary information that would allow their products to use Intel's chips and personal computers. After the acquisition, Intel had about 90,000 employees, including about 12,000 software engineers. In September 2016, Intel sold a majority stake in its computer-security unit to TPG Capital, reversing the five-year-old McAfee acquisition. In August 2010, Intel and Infineon Technologies announced that Intel would acquire Infineon's Wireless Solutions business. Intel planned to use Infineon's technology in laptops, smart phones, netbooks, tablets and embedded computers in consumer products, eventually integrating its wireless modem into Intel's silicon chips. In March 2011, Intel bought most of the assets of Cairo-based SySDSoft. In July 2011, Intel announced that it had agreed to acquire Fulcrum Microsystems Inc., a company specializing in network switches. The company used to be included on the EE Times list of 60 Emerging Startups. In October 2011, Intel reached a deal to acquire Telmap, an Israeli-based navigation software company. The purchase price was not disclosed, but Israeli media reported values around $300 million to $350 million. In July 2012, Intel agreed to buy 10% of the shares of ASML Holding NV for $2.1 billion and another $1 billion for 5% of the shares that need shareholder approval to fund relevant research and development efforts, as part of a EUR3.3 billion ($4.1 billion) deal to accelerate the development of 450-millimeter wafer technology and extreme ultra-violet lithography by as much as two years. In July 2013, Intel confirmed the acquisition of Omek Interactive, an Israeli company that makes technology for gesture-based interfaces, without disclosing the monetary value of the deal. An official statement from Intel read: "The acquisition of Omek Interactive will help increase Intel's capabilities in the delivery of more immersive perceptual computing experiences." One report estimated the value of the acquisition between US$30 million and $50 million. The acquisition of a Spanish natural language recognition startup, Indisys was announced in September 2013. The terms of the deal were not disclosed but an email from an Intel representative stated: "Intel has acquired Indisys, a privately held company based in Seville, Spain. The majority of Indisys employees joined Intel. We signed the agreement to acquire the company on May 31 and the deal has been completed." Indysis explains that its artificial intelligence (AI) technology "is a human image, which converses fluently and with common sense in multiple languages and also works in different platforms". In December 2014, Intel bought PasswordBox. In January 2015, Intel purchased a 30% stake in Vuzix, a smart glasses manufacturer. The deal was worth $24.8 million. In February 2015, Intel announced its agreement to purchase German network chipmaker Lantiq, to aid in its expansion of its range of chips in devices with Internet connection capability. In June 2015, Intel announced its agreement to purchase FPGA design company Altera for $16.7 billion, in its largest acquisition to date. The acquisition completed in December 2015. In October 2015, Intel bought cognitive computing company Saffron Technology for an undisclosed price. In August 2016, Intel purchased deep-learning startup Nervana Systems for over $400 million. In December 2016, Intel acquired computer vision startup Movidius for an undisclosed price. In March 2017, Intel announced that they had agreed to purchase Mobileye, an Israeli developer of "autonomous driving" systems for US$15.3 billion. In June 2017, Intel Corporation announced an investment of over for its upcoming Research and Development (R&D) centre in Bangalore, India. In January 2019, Intel announced an investment of over $11 billion on a new Israeli chip plant, as told by the Israeli Finance Minister. In November 2021, Intel recruited some of the employees of the Centaur Technology division from VIA Technologies, a deal worth $125 million, and effectively acquiring the talent and knowhow of their x86 division. VIA retained the x86 licence and associated patents, and its Zhaoxin CPU joint-venture continues. In December 2021, Intel said it will invest $7.1 billion to build a new chip-packaging and testing factory in Malaysia. The new investment will expand the operations of its Malaysian subsidiary across Penang and Kulim, creating more than 4,000 new Intel jobs and more than 5,000 local construction jobs. In December 2021, Intel announced its plan to take Mobileye automotive unit via an IPO of newly issued stock in 2022, maintaining its majority ownership of the company. In May 2022, Intel announced that they have acquired Finnish graphics technology firm Siru innovations. The firm founded by ex-AMD Qualcomm mobile GPU engineers, is focused on developing software and silicon building blocks for GPU’s made by other companies and is set to join Intel’s fledgling Accelerated Computing Systems and Graphics Group. In May 2022, it was announced that Ericsson and Intel, are pooling research and development to create high-performing Cloud RAN solutions. The organisations have pooled to launch a tech hub in California, USA. The hub focuses on the benefits that Ericsson Cloud RAN and Intel technology can bring to: improving energy efficiency and network performance, reducing time to market, and monetizing new business opportunities such as enterprise applications. Ultrabook fund (2011) In 2011, Intel Capital announced a new fund to support startups working on technologies in line with the company's concept for next generation notebooks. The company is setting aside a $300 million fund to be spent over the next three to four years in areas related to ultrabooks. Intel announced the ultrabook concept at Computex in 2011. The ultrabook is defined as a thin (less than 0.8 inches [~2 cm] thick) notebook that utilizes Intel processors and also incorporates tablet features such as a touch screen and long battery life. At the Intel Developers Forum in 2011, four Taiwan ODMs showed prototype ultrabooks that used Intel's Ivy Bridge chips. Intel plans to improve power consumption of its chips for ultrabooks, like new Ivy Bridge processors in 2013, which will only have 10W default thermal design power. Intel's goal for Ultrabook's price is below $1000; however, according to two presidents from Acer and Compaq, this goal will not be achieved if Intel does not lower the price of its chips. Open source support Intel has a significant participation in the open source communities since 1999. For example, in 2006 Intel released MIT-licensed X.org drivers for their integrated graphic cards of the i965 family of chipsets. Intel released FreeBSD drivers for some networking cards, available under a BSD-compatible license, which were also ported to OpenBSD. Binary firmware files for non-wireless Ethernet devices were also released under a BSD licence allowing free redistribution. Intel ran the Moblin project until April 23, 2009, when they handed the project over to the Linux Foundation. Intel also runs the LessWatts.org campaigns. However, after the release of the wireless products called Intel Pro/Wireless 2100, 2200BG/2225BG/2915ABG and 3945ABG in 2005, Intel was criticized for not granting free redistribution rights for the firmware that must be included in the operating system for the wireless devices to operate. As a result of this, Intel became a target of campaigns to allow free operating systems to include binary firmware on terms acceptable to the open source community. Linspire-Linux creator Michael Robertson outlined the difficult position that Intel was in releasing to open source, as Intel did not want to upset their large customer Microsoft. Theo de Raadt of OpenBSD also claimed that Intel is being "an Open Source fraud" after an Intel employee presented a distorted view of the situation at an open-source conference. In spite of the significant negative attention Intel received as a result of the wireless dealings, the binary firmware still has not gained a license compatible with free software principles. Intel has also supported other open source projects such as Blender and Open 3D Engine. Corporate identity Logo In its history, Intel has had three logos. The first Intel logo featured the company's name stylized in all lowercase, with the letter e dropped below the other letters. The second logo was inspired by the "Intel Inside" campaign, featuring a swirl around the Intel brand name. The third logo, introduced in 2020, was inspired by the previous logos. It removes the swirl as well as the classic blue color in almost all parts of the logo, except for the dot in the "i". Intel Inside Intel has become one of the world's most recognizable computer brands following its long-running Intel Inside campaign. The idea for "Intel Inside" came out of a meeting between Intel and one of the major computer resellers, MicroAge. In the late 1980s, Intel's market share was being seriously eroded by upstart competitors such as Advanced Micro Devices (now AMD), Zilog, and others who had started to sell their less expensive microprocessors to computer manufacturers. This was because, by using cheaper processors, manufacturers could make cheaper computers and gain more market share in an increasingly price-sensitive market. In 1989, Intel's Dennis Carter visited MicroAge's headquarters in Tempe, Arizona, to meet with MicroAge's VP of Marketing, Ron Mion. MicroAge had become one of the largest distributors of Compaq, IBM, HP, and others and thus was a primary although indirect driver of demand for microprocessors. Intel wanted MicroAge to petition its computer suppliers to favor Intel chips. However, Mion felt that the marketplace should decide which processors they wanted. Intel's counterargument was that it would be too difficult to educate PC buyers on why Intel microprocessors were worth paying more for. Mion felt that the public didn't really need to fully understand why Intel chips were better, they just needed to feel they were better. So Mion proposed a market test. Intel would pay for a MicroAge billboard somewhere saying, "If you're buying a personal computer, make sure it has Intel inside." In turn, MicroAge would put "Intel Inside" stickers on the Intel-based computers in their stores in that area. To make the test easier to monitor, Mion decided to do the test in Boulder, Colorado, where it had a single store. Virtually overnight, the sales of personal computers in that store dramatically shifted to Intel-based PCs. Intel very quickly adopted "Intel Inside" as its primary branding and rolled it out worldwide. As is often the case with computer lore, other tidbits have been combined to explain how things evolved. "Intel Inside" has not escaped that tendency and there are other "explanations" that had been floating around. Intel's branding campaign started with "The Computer Inside" tagline in 1990 in the US and Europe. The Japan chapter of Intel proposed an "Intel in it" tagline and kicked off the Japanese campaign by hosting EKI-KON (meaning "Station Concert" in Japanese) at the Tokyo railway station dome on Christmas Day, December 25, 1990. Several months later, "The Computer Inside" incorporated the Japan idea to become "Intel Inside" which eventually elevated to the worldwide branding campaign in 1991, by Intel marketing manager Dennis Carter. A case study, "Inside Intel Inside", was put together by Harvard Business School. The five-note jingle was introduced in 1994 and by its tenth anniversary was being heard in 130 countries around the world. The initial branding agency for the "Intel Inside" campaign was DahlinSmithWhite Advertising of Salt Lake City. The Intel swirl logo was the work of DahlinSmithWhite art director Steve Grigg under the direction of Intel president and CEO Andy Grove. The Intel Inside advertising campaign sought public brand loyalty and awareness of Intel processors in consumer computers. Intel paid some of the advertiser's costs for an ad that used the Intel Inside logo and xylo-marimba jingle. In 2008, Intel planned to shift the emphasis of its Intel Inside campaign from traditional media such as television and print to newer media such as the Internet. Intel required that a minimum of 35% of the money it provided to the companies in its co-op program be used for online marketing. The Intel 2010 annual financial report indicated that $1.8 billion (6% of the gross margin and nearly 16% of the total net income) was allocated to all advertising with Intel Inside being part of that. Intel jingle The D–D–G–D–A xylophone/marimba jingle, known as the "Intel bong", used in Intel advertising was produced by Musikvergnuegen and written by Walter Werzowa, once a member of the Austrian 1980s sampling band Edelweiss. The Intel jingle was made in 1994 to coincide with the launch of the Pentium. It was modified in 1999 to coincide with the launch of the Pentium III, although it overlapped with the 1994 version which was phased out in 2004. Advertisements for products featuring Intel processors with prominent MMX branding featured a version of the jingle with an embellishment (shining sound) after the final note. The jingle was remade a second time in 2004 to coincide with the new logo change. Again, it overlapped with the 1999 version and was not mainstreamed until the launch of the Core processors in 2006, with the melody unchanged. Another remake of the jingle debuted with Intel's new visual identity. The company has made use of numerous variants since its rebranding in 2020 (including the 2004 version). Processor naming strategy In 2006, Intel expanded its promotion of open specification platforms beyond Centrino, to include the Viiv media center PC and the business desktop Intel vPro. In mid-January 2006, Intel announced that they were dropping the long running Pentium name from their processors. The Pentium name was first used to refer to the P5 core Intel processors and was done to comply with court rulings that prevent the trademarking of a string of numbers, so competitors could not just call their processor the same name, as had been done with the prior 386 and 486 processors (both of which had copies manufactured by IBM and AMD). They phased out the Pentium names from mobile processors first, when the new Yonah chips, branded Core Solo and Core Duo, were released. The desktop processors changed when the Core 2 line of processors were released. By 2009, Intel was using a good-better-best strategy with Celeron being good, Pentium better, and the Intel Core family representing the best the company has to offer. According to spokesman Bill Calder, Intel has maintained only the Celeron brand, the Atom brand for netbooks and the vPro lineup for businesses. Since late 2009, Intel's mainstream processors have been called Celeron, Pentium, Core i3, Core i5, Core i7, and Core i9 in order of performance from lowest to highest. The first generation core products carry a 3 digit name, such as i5 750, and the second generation products carry a 4 digit name, such as the i5 2500, and from 10th generation onwards, Intel processors will have a 5 digit name, such as i9 10900K for desktop. In all cases, a 'K' at the end of it shows that it is an unlocked processor, enabling additional overclocking abilities (for instance, 2500K). vPro products will carry the Intel Core i7 vPro processor or the Intel Core i5 vPro processor name. In October 2011, Intel started to sell its Core i7-2700K "Sandy Bridge" chip to customers worldwide. Since 2010, "Centrino" is only being applied to Intel's WiMAX and Wi-Fi technologies. In 2022, Intel announced that they are dropping the Pentium and Celeron naming schemes for their desktop and laptop entry level processors. The "Intel Processor" branding will be replacing the old Pentium and Celeron naming schemes starting in 2023. Typography Neo Sans Intel is a customized version of Neo Sans based on the Neo Sans and Neo Tech, designed by Sebastian Lester in 2004. It was introduced alongside Intel's rebranding in 2006. Previously, Intel used Helvetica as its standard typeface in corporate marketing. Intel Clear is a global font announced in 2014 designed for to be used across all communications. The font family was designed by Red Peek Branding and Dalton Maag Initially available in Latin, Greek and Cyrillic scripts, it replaced Neo Sans Intel as the company's corporate typeface. Intel Clear Hebrew, Intel Clear Arabic were added by Dalton Maag Ltd. Neo Sans Intel remained in logo and to mark processor type and socket on the packaging of Intel's processors. In 2020, as part of a new visual identity, a new typeface, Intel One, was designed. It replaced Intel Clear as the font used by the company in most of its branding, however, it is used alongside Intel Clear typeface. In logo, it replaced Neo Sans Intel typeface. However, it is still used to mark processor type and socket on the packaging of Intel's processors. Intel Brand Book It is a book produced by Red Peak Branding as part of new brand identity campaign, celebrating Intel's achievements while setting the new standard for what Intel looks, feels and sounds like. Litigations and regulatory disputes Patent infringement litigation (2006–2007) In October 2006, a Transmeta lawsuit was filed against Intel for patent infringement on computer architecture and power efficiency technologies. The lawsuit was settled in October 2007, with Intel agreeing to pay US$150 million initially and US$20 million per year for the next five years. Both companies agreed to drop lawsuits against each other, while Intel was granted a perpetual non-exclusive license to use current and future patented Transmeta technologies in its chips for 10 years. Antitrust allegations and litigation (2005–2009) In September 2005, Intel filed a response to an AMD lawsuit, disputing AMD's claims, and claiming that Intel's business practices are fair and lawful. In a rebuttal, Intel deconstructed AMD's offensive strategy and argued that AMD struggled largely as a result of its own bad business decisions, including underinvestment in essential manufacturing capacity and excessive reliance on contracting out chip foundries. Legal analysts predicted the lawsuit would drag on for a number of years, since Intel's initial response indicated its unwillingness to settle with AMD. In 2008, a court date was finally set. On November 4, 2009, New York's attorney general filed an antitrust lawsuit against Intel Corp, claiming the company used "illegal threats and collusion" to dominate the market for computer microprocessors. On November 12, 2009, AMD agreed to drop the antitrust lawsuit against Intel in exchange for $1.25 billion. A joint press release published by the two chip makers stated "While the relationship between the two companies has been difficult in the past, this agreement ends the legal disputes and enables the companies to focus all of our efforts on product innovation and development." An antitrust lawsuit and a class-action suit relating to cold calling employees of other companies has been settled. Allegations by Japan Fair Trade Commission (2005) In 2005, the local Fair Trade Commission found that Intel violated the Japanese Antimonopoly Act. The commission ordered Intel to eliminate discounts that had discriminated against AMD. To avoid a trial, Intel agreed to comply with the order. Allegations by the European Union (2007–2008) In July 2007, the European Commission accused Intel of anti-competitive practices, mostly against AMD. The allegations, going back to 2003, include giving preferential prices to computer makers buying most or all of their chips from Intel, paying computer makers to delay or cancel the launch of products using AMD chips, and providing chips at below standard cost to governments and educational institutions. Intel responded that the allegations were unfounded and instead qualified its market behavior as consumer-friendly. General counsel Bruce Sewell responded that the commission had misunderstood some factual assumptions regarding pricing and manufacturing costs. In February 2008, Intel announced that its office in Munich had been raided by European Union regulators. Intel reported that it was cooperating with investigators. Intel faced a fine of up to 10% of its annual revenue if found guilty of stifling competition. AMD subsequently launched a website promoting these allegations. In June 2008, the EU filed new charges against Intel. In May 2009, the EU found that Intel had engaged in anti-competitive practices and subsequently fined Intel €1.06 billion (US$1.44 billion), a record amount. Intel was found to have paid companies, including Acer, Dell, HP, Lenovo and NEC, to exclusively use Intel chips in their products, and therefore harmed other, less successful companies including AMD. The European Commission said that Intel had deliberately acted to keep competitors out of the computer chip market and in doing so had made a "serious and sustained violation of the EU's antitrust rules". In addition to the fine, Intel was ordered by the commission to immediately cease all illegal practices. Intel has said that they will appeal against the commission's verdict. In June 2014, the General Court, which sits below the European Court of Justice, rejected the appeal. Allegations by regulators in South Korea (2007) In September 2007, South Korean regulators accused Intel of breaking antitrust law. The investigation began in February 2006, when officials raided Intel's South Korean offices. The company risked a penalty of up to 3% of its annual sales if found guilty. In June 2008, the Fair Trade Commission ordered Intel to pay a fine of US$25.5 million for taking advantage of its dominant position to offer incentives to major Korean PC manufacturers on the condition of not buying products from AMD. Allegations by regulators in the United States (2008–2010) New York started an investigation of Intel in January 2008 on whether the company violated antitrust laws in pricing and sales of its microprocessors. In June 2008, the Federal Trade Commission also began an antitrust investigation of the case. In December 2009, the FTC announced it would initiate an administrative proceeding against Intel in September 2010. In November 2009, following a two-year investigation, New York Attorney General Andrew Cuomo sued Intel, accusing them of bribery and coercion, claiming that Intel bribed computer makers to buy more of their chips than those of their rivals and threatened to withdraw these payments if the computer makers were perceived as working too closely with its competitors. Intel has denied these claims. On July 22, 2010, Dell agreed to a settlement with the U.S. Securities and Exchange Commission (SEC) to pay $100M in penalties resulting from charges that Dell did not accurately disclose accounting information to investors. In particular, the SEC charged that from 2002 to 2006, Dell had an agreement with Intel to receive rebates in exchange for not using chips manufactured by AMD. These substantial rebates were not disclosed to investors, but were used to help meet investor expectations regarding the company's financial performance; "These exclusivity payments grew from 10 percent of Dell's operating income in FY 2003 to 38 percent in FY 2006, and peaked at 76 percent in the first quarter of FY 2007." Dell eventually did adopt AMD as a secondary supplier in 2006, and Intel subsequently stopped their rebates, causing Dell's financial performance to fall. Corporate responsibility record Intel has been accused by some residents of Rio Rancho, New Mexico of allowing volatile organic compounds (VOCs) to be released in excess of their pollution permit. One resident claimed that a release of 1.4 tons of carbon tetrachloride was measured from one acid scrubber during the fourth quarter of 2003 but an emission factor allowed Intel to report no carbon tetrachloride emissions for all of 2003. Another resident alleges that Intel was responsible for the release of other VOCs from their Rio Rancho site and that a necropsy of lung tissue from two deceased dogs in the area indicated trace amounts of toluene, hexane, ethylbenzene, and xylene isomers, all of which are solvents used in industrial settings but also commonly found in gasoline, retail paint thinners and retail solvents. During a sub-committee meeting of the New Mexico Environment Improvement Board, a resident claimed that Intel's own reports documented more than of VOCs were released in June and July 2006. Intel's environmental performance is published annually in their corporate responsibility report. Conflict-free production In 2009, Intel announced that it planned to undertake an effort to remove conflict resources—materials sourced from mines whose profits are used to fund armed militant groups, particularly within the Democratic Republic of the Congo—from its supply chain. Intel sought conflict-free sources of the precious metals common to electronics from within the country, using a system of first- and third-party audits, as well as input from the Enough Project and other organizations. During a keynote address at Consumer Electronics Show 2014, Intel CEO at the time, Brian Krzanich, announced that the company's microprocessors would henceforth be conflict free. In 2016, Intel stated that it had expected its entire supply chain to be conflict-free by the end of the year. In its 2012 rankings on the progress of consumer electronics companies relating to conflict minerals, the Enough Project rated Intel the best of 24 companies, calling it a "Pioneer of progress". In 2014, chief executive Brian Krzanich urged the rest of the industry to follow Intel's lead by also shunning conflict minerals. Age discrimination complaints Intel has faced complaints of age discrimination in firing and layoffs. Intel was sued in 1993 by nine former employees, over allegations that they were laid off because they were over the age of 40. A group called FACE Intel (Former and Current Employees of Intel) claims that Intel weeds out older employees. FACE Intel claims that more than 90 percent of people who have been laid off or fired from Intel are over the age of 40. Upside magazine requested data from Intel breaking out its hiring and firing by age, but the company declined to provide any. Intel has denied that age plays any role in Intel's employment practices. FACE Intel was founded by Ken Hamidi, who was fired from Intel in 1995 at the age of 47. Hamidi was blocked in a 1999 court decision from using Intel's email system to distribute criticism of the company to employees, which overturned in 2003 in Intel Corp. v. Hamidi. Tax dispute in India In August 2016, Indian officials of the Bruhat Bengaluru Mahanagara Palike (BBMP) parked garbage trucks on Intel's campus and threatened to dump them for evading payment of property taxes between 2007 and 2008, to the tune of . Intel had reportedly been paying taxes as a non-air-conditioned office, when the campus in fact had central air conditioning. Other factors, such as land acquisition and construction improvements, added to the tax burden. Previously, Intel had appealed the demand in the Karnataka high court in July, during which the court ordered Intel to pay BBMP half the owed amount of plus arrears by August 28 of that year. See also 5 nm process ASCI Red AMD Bumpless Build-up Layer Comparison of ATI Graphics Processing Units Comparison of Intel processors Comparison of Nvidia graphics processing units Cyrix Engineering sample (CPU) Graphics Processing Unit (GPU) Intel Developer Zone (Intel DZ) Intel Driver Update Utility Intel GMA (Graphics Media Accelerator) Intel HD and Iris Graphics Intel Loihi Intel Museum Intel Science Talent Search List of Intel chipsets List of Intel CPU microarchitectures List of Intel manufacturing sites List of Intel microprocessors List of Intel graphics processing units List of mergers and acquisitions by Intel List of semiconductor fabrication plants Wintel Intel Management Engine Intel related biographical articles on Wikipedia Andy Grove Bill Gaede Bob Colwell Craig Barrett (chief executive) Gordon Moore Justin Rattner Pat Gelsinger Paul Otellini Robert Noyce Sean Maloney Notes References External links 1968 establishments in California 1970s initial public offerings American companies established in 1968 Companies based in Santa Clara, California Companies in the Dow Jones Industrial Average Companies listed on the Nasdaq Computer companies established in 1968 Computer companies of the United States Computer memory companies Computer storage companies Foundry semiconductor companies Graphics hardware companies Linux companies Manufacturing companies based in the San Francisco Bay Area Manufacturing companies established in 1968 Mobile phone manufacturers Motherboard companies Multinational companies headquartered in the United States Semiconductor companies of the United States Software companies based in the San Francisco Bay Area Software companies established in 1968 Software companies of the United States Superfund sites in California Technology companies of the United States Technology companies based in the San Francisco Bay Area Technology companies established in 1968
14767
https://en.wikipedia.org/wiki/Economy%20of%20the%20Isle%20of%20Man
Economy of the Isle of Man
The economy of the Isle of Man is a low-tax economy with insurance, online gambling operators and developers, information and communications technology (ICT), and offshore banking forming key sectors of the island's economy. As an offshore financial centre located in the Irish Sea, the Isle of Man is within the British Isles but does not form part of the United Kingdom and is not a member of European Union. As of 2016, the Crown dependency's gross national income (GNI) per capita was US$89,970 as assessed by the World Bank. The Isle of Man Government's own National Income Report shows the largest sectors of the economy are insurance and eGaming with 17% of GNI each, followed by ICT and banking with 9% each, with tourist accommodation in the lowest sector at 0.3%. Economic performance After 32 years of continued Gross Domestic Product (GDP) growth, the financial year 2015/16 showed the first drop in GDP, of 0.9%, triggered by decline in eGaming revenues. The unemployment rate is around 5%. Property prices are flat or declining, but recent figures also show an increase in resident income tax payers. The government's policy of offering incentives to high-technology companies and financial institutions to locate on the island has expanded employment opportunities in high-income industries. Agriculture, fishing, and the hospitality industry, once the mainstays of the economy, now make declining contributions to the island's GNP. The hospitality sector contributed just of 0.3% of GNP in 2015/16, and 629 jobs in 2016. eGaming and ICT contribute the great bulk of GNP. The stability of the island's government and its openness for business make the Isle of Man an attractive alternative jurisdiction (DAW Index ranked 3). Economic strategy In the Vision2020 the Isle of Man government lays out the national strategy of economic growth, seeking an increase of the economically active population an promoting the Island as an 'Enterprise Island, ''Tech Isle', 'Manufacturing centre of excellence', 'Offshore energy hub', 'Destination Island' and for 'Distinctive local food and drink'. The government has published its national economic strategies for several emerging sectors: aerospace, biomed, digital media, ICT. Taxation and trade Tax rates The Isle of Man is a low-tax economy with no capital gains tax, wealth tax, stamp duty, or inheritance tax; and a top rate of income tax of 20%. A tax cap is in force: the maximum amount of tax payable by an individual is £200,000; or £400,000 for couples if they choose to have their incomes jointly assessed. Personal income is assessed and taxed on a total worldwide income basis rather than on a remittance basis. This means that all income earned throughout the world is assessable for Manx tax, rather than only income earned in or brought into the Island. The standard rate of corporation tax for residents and non-residents is 0%; retail business profits above £500,000 and banking business income are taxed at 10%, and rental (or other) income from land and buildings situated on the Isle of Man is taxed at 20%. Trade Trade is mostly with the United Kingdom. The Isle of Man has free access to European Union markets for goods, but only has restricted access for services, people, or financial products. Tax transparency and the offshore banking debate The Isle of Man as an offshore financial centre has been repeatedly featured in the press as a tax haven, most recently in the wake of the Paradise Papers. The Organisation for Economic Co-operation and Development's (OECD) Global Forum on Transparency and Exchange of Information for Tax Purposes has rated the Isle of Man as 'top compliant' for a second time: a status which only three jurisdictions in the world have achieved so far. The island has become the second nation after Austria to ratify a multilateral convention with the OECD to implement measures to prevent Base Erosion and Profit Shifting (BEPS). In a report the European Council lists the Isle of Man together with the other two Crown Dependencies (Guernsey and Jersey) as well as Bermuda, the Cayman Islands and Vanuatu, as committed to addressing the Council's concerns of "Existence of tax regimes that facilitate offshore structures which attract profits without real economic activity" by 2018. Sectors The Isle of Man's Department for Enterprise manages the diversified economy in twelve key sectors. The largest individual sectors by GNI are insurance and eGaming with 17% of GNI each, followed by ICT and banking with 9% each. The 2016 census lists 41,636 total employed. The largest sectors by employment are "medical and health", "financial and business services", construction, retail and public administration. Manufacturing, focused on aerospace and the food and drink industry, employs almost 2000 workers and contributes about 5% of GDP. The sector provides laser optics, industrial diamonds, electronics, plastics and aerospace precision engineering. Finance sector Insurance, banking (includes retail banking, offshore banking and other banking services), other finance and business services, and corporate service providers together contribute the most to the GNI and most of the jobs, with 10,057 people employed in 2016. eGaming & ICT Among the largest employers of the Island's private sector are eGaming (online gambling) companies like The Stars Group, Microgaming, Newfield, and Playtech. The Manx eGaming Association MEGA is representing the sector. Licenses are issued by the Gambling Supervision Commission. In 2005 PokerStars, one of the world's largest online poker sites, relocated its headquarters to the Isle of Man from Costa Rica. In 2006, RNG Gaming a large gaming software developer of P2P tournaments and Get21, a multiplayer online blackjack site, based their corporate offices on the island. The Isle of Man Government Lottery operated from 1986 to 1997. Since 2 December 1999 the island has participated in the United Kingdom National Lottery. The island is the only jurisdiction outside the United Kingdom where it is possible to play the UK National Lottery. Since 2010 it has also been possible for projects in the Isle of Man to receive national lottery Good Causes Funding. The good causes funding is distributed by the Manx Lottery Trust. Tynwald receives the 12p lottery duty for tickets sold in the Island. The shortage of workers with ICT skills is tackled by several initiatives, like an IT and education campus, a new cyber security degree at the University College of Man, a Code Club, and a work permit waiver for skilled immigrants. Filmmaking and digital media Since 1995 Isle of Man Film has co-financed and co-produced over 100 feature film and television dramas which have all filmed on the Island. Among the most successful productions funded in part by Isle of Man Film agency were Waking Ned, where the Manx countryside stood in for rural Ireland, and films like Stormbreaker, Shergar, Tom Brown's Schooldays, I Capture the Castle, The Libertine, Island at War (TV series), Five Children and It, Colour Me Kubrick, Sparkle, and others. Other films that have been filmed on the Isle of Man include Thomas and the Magic Railroad, Harry Potter and the Chamber of Secrets, Keeping Mum and Mindhorn. 2011 Isle of Man Film Oxford Economics was commissioned by Isle of Man Film Ltd to conduct a study into the economic impact of the film industry on the Isle of. Man. The recommendation of this report for Isle of Man Film was to partner with a more established film institution in the UK to source more Isle of Man film production opportunities. This led to the investment of the Isle of Man Government to take shares in Pinewood Shepperton Plc which were sold later with profit. Once one of the busiest areas of film production in the British Isles, the Isle of Man hopes to use its strong foundation in film to grow its television and new digital media industry. In a recent Isle of Man Department of Economic Development strategic review, the Island's over 2,000 jobs counting digital sector features 'digital media' and the creative industries, and embraces partnerships with the industry and its individual sector bodies like the Isle of Media, a new media cluster. Motorsports Hosting of motorsports events, like the Isle of Man Car Rally and the more-prominent TT motorcycle races, contributes to the tourism economy. Tourism Tourism in the Isle of Man developed from advances in transport to the island. In 1819 the first steamship Robert Bruce came to the island, only seven years after the first steam vessel in the UK. In the 1820s, tourism was growing due to improved transport. The island government's own report for the financial years 2014/15-2015/16 shows tourist accommodation to be in the lowest sector at 0.3%, ranking slightly above 'mining and quarrying' (0.1%). Infrastructure Electricity Since 1999, the Isle of Man has received electricity through the world's second longest submarine AC cable, the 90 kV Isle of Man to England Interconnector, as well as from a natural gas power station in Douglas, an oil power station in Peel and a small hydro-electric power station in Sulby Glen. Gas Gas for lighting and heating has been supplied to users on the Isle of Man since 1836, firstly as town gas, then as liquid petroleum gas (LPG); since 2003 natural gas has been available. The future use of hydrogen as a supplementary or substitute fuel is being studied. Broadband The Island is connected with five submarine cables to the UK and Ireland. While the Isle of Man Communications Commission refers to Akamai’s recent State of the Internet Report for Q1 2017, with "the Island ranked 8th in the world for percentage of broadband connections with >4 Mb/s connectivity, with 96% of users connecting at speeds greater than 4Mb/s", an "international league table of broadband speeds puts the Isle of Man at 50th in the world". Manx Telecom recently announced to roll out Fibre-to-the-Home (FTTH) superfast broadband with download speeds of up to 1Gigabit per second. Travel links Ronaldsway Airport links the Isle of Man with six airlines to eleven UK and Irish scheduled flight destinations. The Steam Packet Company provides ferry services to Liverpool, Heysham, Belfast and Dublin. Statistics Labour force—by occupation: agriculture, forestry and fishing 3%, manufacturing 11%, construction 10%, transport and communication 8%, wholesale and retail distribution 11%, professional and scientific services 18%, public administration 6%, banking and finance 18%, tourism 2%, entertainment and catering 3%, miscellaneous services 10% Unemployment rate: nominally 5.0% (July 2020) Industries: financial services, light manufacturing, tourism Agriculture—products: cereals, vegetables, cattle, sheep, pigs, poultry Exports: $NA Exports—commodities: tweeds, herring, processed shellfish, beef, lamb Exports—partners: UK Imports: $NA Imports—commodities: timber, fertilizers, fish Imports—partners: UK Debt—external: $NA Economic aid—recipient: $NA Currency: 1 Isle of Man pound = 100 pence Exchange rates: the Manx pound is at par with the British pound Fiscal year: 1 April – 31 March See also Economy of Europe Notes References Isle of Man
14946
https://en.wikipedia.org/wiki/Ice
Ice
Ice is water frozen into a solid state, typically forming at or below temperatures of 32 °F, 0 °C, or 273.15 K. Depending on the presence of impurities such as particles of soil or bubbles of air, it can appear transparent or a more or less opaque bluish-white color. In the Solar System, ice is abundant and occurs naturally from as close to the Sun as Mercury to as far away as the Oort cloud objects. Beyond the Solar System, it occurs as interstellar ice. It is abundant on Earth's surfaceparticularly in the polar regions and above the snow lineand, as a common form of precipitation and deposition, plays a key role in Earth's water cycle and climate. It falls as snowflakes and hail or occurs as frost, icicles or ice spikes and aggregates from snow as glaciers and ice sheets. Ice exhibits at least eighteen phases (packing geometries), depending on temperature and pressure. When water is cooled rapidly (quenching), up to three types of amorphous ice can form depending on its history of pressure and temperature. When cooled slowly, correlated proton tunneling occurs below (, ) giving rise to macroscopic quantum phenomena. Virtually all ice on Earth's surface and in its atmosphere is of a hexagonal crystalline structure denoted as ice Ih (spoken as "ice one h") with minute traces of cubic ice, denoted as ice Ic and, more recently found, Ice VII inclusions in diamonds. The most common phase transition to ice Ih occurs when liquid water is cooled below (, ) at standard atmospheric pressure. It may also be deposited directly by water vapor, as happens in the formation of frost. The transition from ice to water is melting and from ice directly to water vapor is sublimation. Ice is used in a variety of ways, including for cooling, for winter sports, and ice sculpting. Physical properties As a naturally occurring crystalline inorganic solid with an ordered structure, ice is considered to be a mineral. It possesses a regular crystalline structure based on the molecule of water, which consists of a single oxygen atom covalently bonded to two hydrogen atoms, or H–O–H. However, many of the physical properties of water and ice are controlled by the formation of hydrogen bonds between adjacent oxygen and hydrogen atoms; while it is a weak bond, it is nonetheless critical in controlling the structure of both water and ice. An unusual property of water is that its solid form—ice frozen at atmospheric pressure—is approximately 8.3% less dense than its liquid form; this is equivalent to a volumetric expansion of 9%. The density of ice is 0.9167–0.9168 g/cm3 at 0 °C and standard atmospheric pressure (101,325 Pa), whereas water has a density of 0.9998–0.999863 g/cm3 at the same temperature and pressure. Liquid water is densest, essentially 1.00 g/cm3, at 4 °C and begins to lose its density as the water molecules begin to form the hexagonal crystals of ice as the freezing point is reached. This is due to hydrogen bonding dominating the intermolecular forces, which results in a packing of molecules less compact in the solid. Density of ice increases slightly with decreasing temperature and has a value of 0.9340 g/cm3 at −180 °C (93 K). When water freezes, it increases in volume (about 9% for fresh water). The effect of expansion during freezing can be dramatic, and ice expansion is a basic cause of freeze-thaw weathering of rock in nature and damage to building foundations and roadways from frost heaving. It is also a common cause of the flooding of houses when water pipes burst due to the pressure of expanding water when it freezes. The result of this process is that ice (in its most common form) floats on liquid water, which is an important feature in Earth's biosphere. It has been argued that without this property, natural bodies of water would freeze, in some cases permanently, from the bottom up, resulting in a loss of bottom-dependent animal and plant life in fresh and sea water. Sufficiently thin ice sheets allow light to pass through while protecting the underside from short-term weather extremes such as wind chill. This creates a sheltered environment for bacterial and algal colonies. When sea water freezes, the ice is riddled with brine-filled channels which sustain sympagic organisms such as bacteria, algae, copepods and annelids, which in turn provide food for animals such as krill and specialised fish like the bald notothen, fed upon in turn by larger animals such as emperor penguins and minke whales. When ice melts, it absorbs as much energy as it would take to heat an equivalent mass of water by 80 °C. During the melting process, the temperature remains constant at 0 °C. While melting, any energy added breaks the hydrogen bonds between ice (water) molecules. Energy becomes available to increase the thermal energy (temperature) only after enough hydrogen bonds are broken that the ice can be considered liquid water. The amount of energy consumed in breaking hydrogen bonds in the transition from ice to water is known as the heat of fusion. As with water, ice absorbs light at the red end of the spectrum preferentially as the result of an overtone of an oxygen–hydrogen (O–H) bond stretch. Compared with water, this absorption is shifted toward slightly lower energies. Thus, ice appears blue, with a slightly greener tint than liquid water. Since absorption is cumulative, the color effect intensifies with increasing thickness or if internal reflections cause the light to take a longer path through the ice. Other colors can appear in the presence of light absorbing impurities, where the impurity is dictating the color rather than the ice itself. For instance, icebergs containing impurities (e.g., sediments, algae, air bubbles) can appear brown, grey or green. Because ice in natural environments is usually close to its melting temperature, its hardness shows pronounced temperature variations. At its melting point, ice has a Mohs hardness of 2 or less, but the hardness increases to about 4 at a temperature of and to 6 at a temperature of , the vaporization point of solid carbon dioxide (dry ice). Phases Ice may be any one of the 19 known solid crystalline phases of water, or in an amorphous solid state at various densities. Most liquids under increased pressure freeze at higher temperatures because the pressure helps to hold the molecules together. However, the strong hydrogen bonds in water make it different: for some pressures higher than , water freezes at a temperature below 0 °C, as shown in the phase diagram below. The melting of ice under high pressures is thought to contribute to the movement of glaciers. Ice, water, and water vapour can coexist at the triple point, which is exactly at a pressure of 611.657 Pa. The kelvin was defined as of the difference between this triple point and absolute zero, though this definition changed in May 2019. Unlike most other solids, ice is difficult to superheat. In an experiment, ice at −3 °C was superheated to about 17 °C for about 250 picoseconds. Subjected to higher pressures and varying temperatures, ice can form in 19 separate known crystalline phases. With care, at least 15 of these phases (one of the known exceptions being ice X) can be recovered at ambient pressure and low temperature in metastable form. The types are differentiated by their crystalline structure, proton ordering, and density. There are also two metastable phases of ice under pressure, both fully hydrogen-disordered; these are IV and XII. Ice XII was discovered in 1996. In 2006, XIII and XIV were discovered. Ices XI, XIII, and XIV are hydrogen-ordered forms of ices I, V, and XII respectively. In 2009, ice XV was found at extremely high pressures and −143 °C. At even higher pressures, ice is predicted to become a metal; this has been variously estimated to occur at 1.55 TPa or 5.62 TPa. As well as crystalline forms, solid water can exist in amorphous states as amorphous solid water (ASW) of varying densities. Water in the interstellar medium is dominated by amorphous ice, making it likely the most common form of water in the universe. Low-density ASW (LDA), also known as hyperquenched glassy water, may be responsible for noctilucent clouds on Earth and is usually formed by deposition of water vapor in cold or vacuum conditions. High-density ASW (HDA) is formed by compression of ordinary ice I or LDA at GPa pressures. Very-high-density ASW (VHDA) is HDA slightly warmed to 160 K under 1–2 GPa pressures. In outer space, hexagonal crystalline ice (the predominant form found on Earth) is extremely rare. Amorphous ice is more common; however, hexagonal crystalline ice can be formed by volcanic action. Ice from a theorized superionic water may possess two crystalline structures. At pressures in excess of such superionic ice would take on a body-centered cubic structure. However, at pressures in excess of the structure may shift to a more stable face-centered cubic lattice. It is speculated that superionic ice could compose the interior of ice giants such as Uranus and Neptune. Friction properties The low coefficient of friction ("slipperiness") of ice has been attributed to the pressure of an object coming into contact with the ice, melting a thin layer of the ice and allowing the object to glide across the surface. For example, the blade of an ice skate, upon exerting pressure on the ice, would melt a thin layer, providing lubrication between the ice and the blade. This explanation, called "pressure melting", originated in the 19th century. However, it does not account for skating on ice temperatures lower than , which is often skated upon. Also, the effect of pressure melting is too small to account for the reduced friction as commonly experienced. A second theory describing the coefficient of friction of ice suggested that ice molecules at the interface cannot properly bond with the molecules of the mass of ice beneath (and thus are free to move like molecules of liquid water). These molecules remain in a semi-liquid state, providing lubrication regardless of pressure against the ice exerted by any object. However, the significance of this hypothesis is disputed by experiments showing a high coefficient of friction for ice using atomic force microscopy. A third theory is "friction heating", which suggests that friction of the material is the cause of the ice layer melting. However, this theory does not sufficiently explain why ice is slippery when standing still even at below-zero temperatures. A comprehensive theory of ice friction takes into account all the above-mentioned friction mechanisms. This model allows quantitative estimation of the friction coefficient of ice against various materials as a function of temperature and sliding speed. In typical conditions related to winter sports and tires of a vehicle on ice, melting of a thin ice layer due to the frictional heating is the primary reason for the slipperiness. The mechanism controlling the frictional properties of ice is still an active area of scientific study. Natural formation The term that collectively describes all of the parts of the Earth's surface where water is in frozen form is the cryosphere. Ice is an important component of the global climate, particularly in regard to the water cycle. Glaciers and snowpacks are an important storage mechanism for fresh water; over time, they may sublimate or melt. Snowmelt is an important source of seasonal fresh water. The World Meteorological Organization defines several kinds of ice depending on origin, size, shape, influence and so on. Clathrate hydrates are forms of ice that contain gas molecules trapped within its crystal lattice. On the oceans Ice that is found at sea may be in the form of drift ice floating in the water, fast ice fixed to a shoreline or anchor ice if attached to the sea bottom. Ice which calves (breaks off) from an ice shelf or glacier may become an iceberg. Sea ice can be forced together by currents and winds to form pressure ridges up to tall. Navigation through areas of sea ice occurs in openings called "polynyas" or "leads" or requires the use of a special ship called an "icebreaker". On land and structures Ice on land ranges from the largest type called an "ice sheet" to smaller ice caps and ice fields to glaciers and ice streams to the snow line and snow fields. Aufeis is layered ice that forms in Arctic and subarctic stream valleys. Ice, frozen in the stream bed, blocks normal groundwater discharge, and causes the local water table to rise, resulting in water discharge on top of the frozen layer. This water then freezes, causing the water table to rise further and repeat the cycle. The result is a stratified ice deposit, often several meters thick. Freezing rain is a type of winter storm called an ice storm where rain falls and then freezes producing a glaze of ice. Ice can also form icicles, similar to stalactites in appearance, or stalagmite-like forms as water drips and re-freezes. The term "ice dam" has three meanings (others discussed below). On structures, an ice dam is the buildup of ice on a sloped roof which stops melt water from draining properly and can cause damage from water leaks in buildings. On rivers and streams Ice which forms on moving water tends to be less uniform and stable than ice which forms on calm water. Ice jams (sometimes called "ice dams"), when broken chunks of ice pile up, are the greatest ice hazard on rivers. Ice jams can cause flooding, damage structures in or near the river, and damage vessels on the river. Ice jams can cause some hydropower industrial facilities to completely shut down. An ice dam is a blockage from the movement of a glacier which may produce a proglacial lake. Heavy ice flows in rivers can also damage vessels and require the use of an icebreaker to keep navigation possible. Ice discs are circular formations of ice surrounded by water in a river. Pancake ice is a formation of ice generally created in areas with less calm conditions. On lakes Ice forms on calm water from the shores, a thin layer spreading across the surface, and then downward. Ice on lakes is generally four types: primary, secondary, superimposed and agglomerate. Primary ice forms first. Secondary ice forms below the primary ice in a direction parallel to the direction of the heat flow. Superimposed ice forms on top of the ice surface from rain or water which seeps up through cracks in the ice which often settles when loaded with snow. Shelf ice occurs when floating pieces of ice are driven by the wind piling up on the windward shore. Candle ice is a form of rotten ice that develops in columns perpendicular to the surface of a lake. An ice shove occurs when ice movement, caused by ice expansion and/or wind action, occurs to the extent that ice pushes onto the shores of lakes, often displacing sediment that makes up the shoreline. In the air Rime Rime is a type of ice formed on cold objects when drops of water crystallize on them. This can be observed in foggy weather, when the temperature drops during the night. Soft rime contains a high proportion of trapped air, making it appear white rather than transparent, and giving it a density about one quarter of that of pure ice. Hard rime is comparatively dense. Pellets Ice pellets are a form of precipitation consisting of small, translucent balls of ice. This form of precipitation is also referred to as "sleet" by the United States National Weather Service. (In British English "sleet" refers to a mixture of rain and snow.) Ice pellets are usually smaller than hailstones. They often bounce when they hit the ground, and generally do not freeze into a solid mass unless mixed with freezing rain. The METAR code for ice pellets is PL. Ice pellets form when a layer of above-freezing air is located between above the ground, with sub-freezing air both above and below it. This causes the partial or complete melting of any snowflakes falling through the warm layer. As they fall back into the sub-freezing layer closer to the surface, they re-freeze into ice pellets. However, if the sub-freezing layer beneath the warm layer is too small, the precipitation will not have time to re-freeze, and freezing rain will be the result at the surface. A temperature profile showing a warm layer above the ground is most likely to be found in advance of a warm front during the cold season, but can occasionally be found behind a passing cold front. Hail Like other precipitation, hail forms in storm clouds when supercooled water droplets freeze on contact with condensation nuclei, such as dust or dirt. The storm's updraft blows the hailstones to the upper part of the cloud. The updraft dissipates and the hailstones fall down, back into the updraft, and are lifted up again. Hail has a diameter of or more. Within METAR code, GR is used to indicate larger hail, of a diameter of at least and GS for smaller. Stones of , and are the most frequently reported hail sizes in North America. Hailstones can grow to and weigh more than . In large hailstones, latent heat released by further freezing may melt the outer shell of the hailstone. The hailstone then may undergo 'wet growth', where the liquid outer shell collects other smaller hailstones. The hailstone gains an ice layer and grows increasingly larger with each ascent. Once a hailstone becomes too heavy to be supported by the storm's updraft, it falls from the cloud. Hail forms in strong thunderstorm clouds, particularly those with intense updrafts, high liquid water content, great vertical extent, large water droplets, and where a good portion of the cloud layer is below freezing . Hail-producing clouds are often identifiable by their green coloration. The growth rate is maximized at about , and becomes vanishingly small much below as supercooled water droplets become rare. For this reason, hail is most common within continental interiors of the mid-latitudes, as hail formation is considerably more likely when the freezing level is below the altitude of . Entrainment of dry air into strong thunderstorms over continents can increase the frequency of hail by promoting evaporational cooling which lowers the freezing level of thunderstorm clouds giving hail a larger volume to grow in. Accordingly, hail is actually less common in the tropics despite a much higher frequency of thunderstorms than in the mid-latitudes because the atmosphere over the tropics tends to be warmer over a much greater depth. Hail in the tropics occurs mainly at higher elevations. Snow Snow crystals form when tiny supercooled cloud droplets (about 10 μm in diameter) freeze. These droplets are able to remain liquid at temperatures lower than , because to freeze, a few molecules in the droplet need to get together by chance to form an arrangement similar to that in an ice lattice; then the droplet freezes around this "nucleus". Experiments show that this "homogeneous" nucleation of cloud droplets only occurs at temperatures lower than . In warmer clouds an aerosol particle or "ice nucleus" must be present in (or in contact with) the droplet to act as a nucleus. Our understanding of what particles make efficient ice nuclei is poor – what we do know is they are very rare compared to that cloud condensation nuclei on which liquid droplets form. Clays, desert dust and biological particles may be effective, although to what extent is unclear. Artificial nuclei are used in cloud seeding. The droplet then grows by condensation of water vapor onto the ice surfaces. Diamond dust So-called "diamond dust", also known as ice needles or ice crystals, forms at temperatures approaching due to air with slightly higher moisture from aloft mixing with colder, surface-based air. The METAR identifier for diamond dust within international hourly weather reports is IC. Ablation Ablation of ice refers to both its melting and its dissolution. The melting of ice means entails the breaking of hydrogen bonds between the water molecules. The ordering of the molecules in the solid breaks down to a less ordered state and the solid melts to become a liquid. This is achieved by increasing the internal energy of the ice beyond the melting point. When ice melts it absorbs as much energy as would be required to heat an equivalent amount of water by 80 °C. While melting, the temperature of the ice surface remains constant at 0 °C. The rate of the melting process depends on the efficiency of the energy exchange process. An ice surface in fresh water melts solely by free convection with a rate that depends linearly on the water temperature, T∞, when T∞ is less than 3.98 °C, and superlinearly when T∞ is equal to or greater than 3.98 °C, with the rate being proportional to (T∞ − 3.98 °C)α, with α =  for T∞ much greater than 8 °C, and α =  for in between temperatures T∞. In salty ambient conditions, dissolution rather than melting often causes the ablation of ice. For example, the temperature of the Arctic Ocean is generally below the melting point of ablating sea ice. The phase transition from solid to liquid is achieved by mixing salt and water molecules, similar to the dissolution of sugar in water, even though the water temperature is far below the melting point of the sugar. Thus the dissolution rate is limited by salt transport whereas melting can occur at much higher rates that are characteristic for heat transport. Role in human activities Humans have used ice for cooling and food preservation for centuries, relying on harvesting natural ice in various forms and then transitioning to the mechanical production of the material. Ice also presents a challenge to transportation in various forms and a setting for winter sports. Cooling Ice has long been valued as a means of cooling. In 400 BC Iran, Persian engineers had already mastered the technique of storing ice in the middle of summer in the desert. The ice was brought in from ice pools or during the winters from nearby mountains in bulk amounts, and stored in specially designed, naturally cooled refrigerators, called yakhchal (meaning ice storage). This was a large underground space (up to 5000 m3) that had thick walls (at least two meters at the base) made of a special mortar called sarooj, composed of sand, clay, egg whites, lime, goat hair, and ash in specific proportions, and which was known to be resistant to heat transfer. This mixture was thought to be completely water impenetrable. The space often had access to a qanat, and often contained a system of windcatchers which could easily bring temperatures inside the space down to frigid levels on summer days. The ice was used to chill treats for royalty. Harvesting There were thriving industries in 16th–17th century England whereby low-lying areas along the Thames Estuary were flooded during the winter, and ice harvested in carts and stored inter-seasonally in insulated wooden houses as a provision to an icehouse often located in large country houses, and widely used to keep fish fresh when caught in distant waters. This was allegedly copied by an Englishman who had seen the same activity in China. Ice was imported into England from Norway on a considerable scale as early as 1823. In the United States, the first cargo of ice was sent from New York City to Charleston, South Carolina, in 1799, and by the first half of the 19th century, ice harvesting had become a big business. Frederic Tudor, who became known as the "Ice King", worked on developing better insulation products for long distance shipments of ice, especially to the tropics; this became known as the ice trade. Trieste sent ice to Egypt, Corfu, and Zante; Switzerland, to France; and Germany sometimes was supplied from Bavarian lakes. The Hungarian Parliament building used ice harvested in the winter from Lake Balaton for air conditioning. Ice houses were used to store ice formed in the winter, to make ice available all year long, and an early type of refrigerator known as an icebox was cooled using a block of ice placed inside it. In many cities, it was not unusual to have a regular ice delivery service during the summer. The advent of artificial refrigeration technology has since made delivery of ice obsolete. Ice is still harvested for ice and snow sculpture events. For example, a swing saw is used to get ice for the Harbin International Ice and Snow Sculpture Festival each year from the frozen surface of the Songhua River. Artificial production The earliest known written process to artificially make ice is by the 13th-century writings of Arab historian Ibn Abu Usaybia in his book Kitab Uyun al-anba fi tabaqat-al-atibba concerning medicine in which Ibn Abu Usaybi’a attributes the process to an even older author, Ibn Bakhtawayhi, of whom nothing is known. Mechanical production Ice is now produced on an industrial scale, for uses including food storage and processing, chemical manufacturing, concrete mixing and curing, and consumer or packaged ice. Most commercial icemakers produce three basic types of fragmentary ice: flake, tubular and plate, using a variety of techniques. Large batch ice makers can produce up to 75 tons of ice per day. In 2002, there were 426 commercial ice-making companies in the United States, with a combined value of shipments of $595,487,000. Home refrigerators can also make ice with a built in icemaker, which will typically make ice cubes or crushed ice. Stand-alone icemaker units that make ice cubes are often called ice machines. Transportation Ice can present challenges to safe transportation on land, sea and in the air. Land travel Ice forming on roads is a dangerous winter hazard. Black ice is very difficult to see, because it lacks the expected frosty surface. Whenever there is freezing rain or snow which occurs at a temperature near the melting point, it is common for ice to build up on the windows of vehicles. Driving safely requires the removal of the ice build-up. Ice scrapers are tools designed to break the ice free and clear the windows, though removing the ice can be a long and laborious process. Far enough below the freezing point, a thin layer of ice crystals can form on the inside surface of windows. This usually happens when a vehicle has been left alone after being driven for a while, but can happen while driving, if the outside temperature is low enough. Moisture from the driver's breath is the source of water for the crystals. It is troublesome to remove this form of ice, so people often open their windows slightly when the vehicle is parked in order to let the moisture dissipate, and it is now common for cars to have rear-window defrosters to solve the problem. A similar problem can happen in homes, which is one reason why many colder regions require double-pane windows for insulation. When the outdoor temperature stays below freezing for extended periods, very thick layers of ice can form on lakes and other bodies of water, although places with flowing water require much colder temperatures. The ice can become thick enough to drive onto with automobiles and trucks. Doing this safely requires a thickness of at least . Water-borne travel For ships, ice presents two distinct hazards. First, spray and freezing rain can produce an ice build-up on the superstructure of a vessel sufficient to make it unstable, and to require it to be hacked off or melted with steam hoses. Second, icebergs – large masses of ice floating in water (typically created when glaciers reach the sea) – can be dangerous if struck by a ship when underway. Icebergs have been responsible for the sinking of many ships, the most famous being the Titanic. For harbors near the poles, being ice-free, ideally all year long, is an important advantage. Examples are Murmansk (Russia), Petsamo (Russia, formerly Finland), and Vardø (Norway). Harbors which are not ice-free are opened up using icebreakers. Air travel For aircraft, ice can cause a number of dangers. As an aircraft climbs, it passes through air layers of different temperature and humidity, some of which may be conducive to ice formation. If ice forms on the wings or control surfaces, this may adversely affect the flying qualities of the aircraft. During the first non-stop flight across the Atlantic, the British aviators Captain John Alcock and Lieutenant Arthur Whitten Brown encountered such icing conditions – Brown left the cockpit and climbed onto the wing several times to remove ice which was covering the engine air intakes of the Vickers Vimy aircraft they were flying. One vulnerability effected by icing that is associated with reciprocating internal combustion engines is the carburetor. As air is sucked through the carburetor into the engine, the local air pressure is lowered, which causes adiabatic cooling. Thus, in humid near-freezing conditions, the carburetor will be colder, and tend to ice up. This will block the supply of air to the engine, and cause it to fail. For this reason, aircraft reciprocating engines with carburetors are provided with carburetor air intake heaters. The increasing use of fuel injection—which does not require carburetors—has made "carb icing" less of an issue for reciprocating engines. Jet engines do not experience carb icing, but recent evidence indicates that they can be slowed, stopped, or damaged by internal icing in certain types of atmospheric conditions much more easily than previously believed. In most cases, the engines can be quickly restarted and flights are not endangered, but research continues to determine the exact conditions which produce this type of icing, and find the best methods to prevent, or reverse it, in flight. Recreation and sports Ice also plays a central role in winter recreation and in many sports such as ice skating, tour skating, ice hockey, bandy, ice fishing, ice climbing, curling, broomball and sled racing on bobsled, luge and skeleton. Many of the different sports played on ice get international attention every four years during the Winter Olympic Games. A sort of sailboat on blades gives rise to ice yachting. Another sport is ice racing, where drivers must speed on lake ice, while also controlling the skid of their vehicle (similar in some ways to dirt track racing). The sport has even been modified for ice rinks. Other uses As thermal ballast Ice is used to cool and preserve food in iceboxes. Ice cubes or crushed ice can be used to cool drinks. As the ice melts, it absorbs heat and keeps the drink near . Ice can be used as part of an air conditioning system, using battery- or solar-powered fans to blow hot air over the ice. This is especially useful during heat waves when power is out and standard (electrically powered) air conditioners do not work. Ice can be used (like other cold packs) to reduce swelling (by decreasing blood flow) and pain by pressing it against an area of the body. As structural material Engineers used the substantial strength of pack ice when they constructed Antarctica's first floating ice pier in 1973. Such ice piers are used during cargo operations to load and offload ships. Fleet operations personnel make the floating pier during the winter. They build upon naturally occurring frozen seawater in McMurdo Sound until the dock reaches a depth of about . Ice piers have a lifespan of three to five years. Structures and ice sculptures are built out of large chunks of ice or by spraying water The structures are mostly ornamental (as in the case with ice castles), and not practical for long-term habitation. Ice hotels exist on a seasonal basis in a few cold areas. Igloos are another example of a temporary structure, made primarily from snow. In cold climates, roads are regularly prepared on iced-over lakes and archipelago areas. Temporarily, even a railroad has been built on ice. During World War II, Project Habbakuk was an Allied programme which investigated the use of pykrete (wood fibers mixed with ice) as a possible material for warships, especially aircraft carriers, due to the ease with which a vessel immune to torpedoes, and a large deck, could be constructed by ice. A small-scale prototype was built, but the need for such a vessel in the war was removed prior to building it in full-scale. Ice has even been used as the material for a variety of musical instruments, for example by percussionist Terje Isungset. Non-water The solid phases of several other volatile substances are also referred to as ices; generally a volatile is classed as an ice if its melting point lies above or around 100 K. The best known example is dry ice, the solid form of carbon dioxide. A "magnetic analogue" of ice is also realized in some insulating magnetic materials in which the magnetic moments mimic the position of protons in water ice and obey energetic constraints similar to the Bernal-Fowler ice rules arising from the geometrical frustration of the proton configuration in water ice. These materials are called spin ice. See also References External links Webmineral listing for Ice MinDat.org listing and location data for Ice Estimating the bearing capacity of ice High-temperature, high-pressure ice The Surprisingly Cool History of Ice Glaciology Minerals Transparent materials Articles containing video clips Limnology Oceanography Cryosphere
15085
https://en.wikipedia.org/wiki/Iconoclasm
Iconoclasm
Iconoclasm (from Greek: + ) is the social belief in the importance of the destruction of icons and other images or monuments, most frequently for religious or political reasons. People who engage in or support iconoclasm are called iconoclasts, a term that has come to be figuratively applied to any individual who challenges "cherished beliefs or venerated institutions on the grounds that they are erroneous or pernicious." Conversely, one who reveres or venerates religious images is called (by iconoclasts) an iconolater; in a Byzantine context, such a person is called an iconodule or iconophile. Iconoclasm does not generally encompass the destruction of the images of a specific ruler after his or her death or overthrow, a practice better known as damnatio memoriae. While iconoclasm may be carried out by adherents of a different religion, it is more commonly the result of sectarian disputes between factions of the same religion. The term originates from the Byzantine Iconoclasm, the struggles between proponents and opponents of religious icons in the Byzantine Empire from 726 to 842 AD. Degrees of iconoclasm vary greatly among religions and their branches, but are strongest in religions which oppose idolatry, including the Abrahamic religions. Outside of the religious context, iconoclasm can refer to movements for widespread destruction in symbols of an ideology or cause, such as the destruction of monarchist symbols during the French Revolution. Early religious iconoclasm Ancient era In the Bronze Age, the most significant episode of iconoclasm occurred in Egypt during the Amarna Period, when Akhenaten, based in his new capital of Akhetaten, instituted a significant shift in Egyptian artistic styles alongside a campaign of intolerance towards the traditional gods and a new emphasis on a state monolatristic tradition focused on the god Aten, the Sun disk—many temples and monuments were destroyed as a result: In rebellion against the old religion and the powerful priests of Amun, Akhenaten ordered the eradication of all of Egypt's traditional gods. He sent royal officials to chisel out and destroy every reference to Amun and the names of other deities on tombs, temple walls, and cartouches to instill in the people that the Aten was the one true god. Public references to Akhenaten were destroyed soon after his death. Comparing the ancient Egyptians with the Israelites, Jan Assmann writes: For Egypt, the greatest horror was the destruction or abduction of the cult images. In the eyes of the Israelites, the erection of images meant the destruction of divine presence; in the eyes of the Egyptians, this same effect was attained by the destruction of images. In Egypt, iconoclasm was the most terrible religious crime; in Israel, the most terrible religious crime was idolatry. In this respect Osarseph alias Akhenaten, the iconoclast, and the Golden Calf, the paragon of idolatry, correspond to each other inversely, and it is strange that Aaron could so easily avoid the role of the religious criminal. It is more than probable that these traditions evolved under mutual influence. In this respect, Moses and Akhenaten became, after all, closely related. Judaism According to the Hebrew Bible, God instructed the Israelites to "destroy all [the] engraved stones, destroy all [the] molded images, and demolish all [the] high places" of the indigenous Canaanite population as soon as they entered the Promised Land. In Judaism, King Hezekiah purged Solomon's Temple in Jerusalem and all figures were also destroyed in the Land of Israel, including the Nehushtan, as recorded in the Second Book of Kings. His reforms were reversed in the reign of his son Manasseh. Iconoclasm in Christian history Scattered expressions of opposition to the use of images have been reported: the Synod of Elvira appeared to endorse iconoclasm; Canon 36 states, "Pictures are not to be placed in churches, so that they do not become objects of worship and adoration." A possible translation is also: "There shall be no pictures in the church, lest what is worshipped and adored should be depicted on the walls." The date of this canon is disputed. Proscription ceased after the destruction of pagan temples. However, widespread use of Christian iconography only began as Christianity increasingly spread among Gentiles after the legalization of Christianity by Roman Emperor Constantine (c. 312 AD). During the process of Christianisation under Constantine, Christian groups destroyed the images and sculptures expressive of the Roman Empire's polytheist state religion. Among early church theologians, iconoclastic tendencies were supported by theologians such as: Tertullian, Clement of Alexandria, Origen, Lactantius, Justin Martyr, Eusebius and Epiphanus. Byzantine era The period after the reign of Byzantine Emperor Justinian (527–565) evidently saw a huge increase in the use of images, both in volume and quality, and a gathering aniconic reaction. One notable change within the Byzantine Empire came in 695, when Justinian II's government added a full-face image of Christ on the obverse of imperial gold coins. The change caused the Caliph Abd al-Malik to stop his earlier adoption of Byzantine coin types. He started a purely Islamic coinage with lettering only. A letter by the Patriarch Germanus, written before 726 to two iconoclast bishops, says that "now whole towns and multitudes of people are in considerable agitation over this matter," but there is little written evidence of the debate. Government-led iconoclasm began with Byzantine Emperor Leo III, who issued a series of edicts between 726 and 730 against the veneration of images. The religious conflict created political and economic divisions in Byzantine society; iconoclasm was generally supported by the Eastern, poorer, non-Greek peoples of the Empire who had to frequently deal with raids from the new Muslim Empire. On the other hand, the wealthier Greeks of Constantinople and the peoples of the Balkan and Italian provinces strongly opposed iconoclasm. Pre-Reformation Peter of Bruys opposed the usage of religious images, the Strigolniki were also possibly iconoclastic. Claudius of Turin was the bishop of Turin from 817 until his death. He is most noted for teaching iconoclasm. Reformation era The first iconoclastic wave happened in Wittenberg in the early 1520s under reformers Thomas Müntzer and Andreas Karlstadt, in the absence of Martin Luther, who then, concealed under the pen-name of 'Junker Jörg', intervened to calm things down. Luther argued that the mental picturing of Christ when reading the Scriptures was similar in character to artistic renderings of Christ. In contrast to the Lutherans who favoured certain types of sacred art in their churches and homes, the Reformed (Calvinist) leaders, in particular Andreas Karlstadt, Huldrych Zwingli and John Calvin, encouraged the removal of religious images by invoking the Decalogue's prohibition of idolatry and the manufacture of graven (sculpted) images of God. As a result, individuals attacked statues and images, most famously in the beeldenstorm across the Low Countries in 1566. However, in most cases, civil authorities removed images in an orderly manner in the newly Reformed Protestant cities and territories of Europe. The belief of iconoclasm caused havoc throughout Europe. In 1523, specifically due to the Swiss reformer Huldrych Zwingli, a vast number of his followers viewed themselves as being involved in a spiritual community that in matters of faith should obey neither the visible Church nor lay authorities. According to Peter George Wallace "Zwingli's attack on images, at the first debate, triggered iconoclastic incidents in Zurich and the villages under civic jurisdiction that the reformer was unwilling to condone." Due to this action of protest against authority, "Zwingli responded with a carefully reasoned treatise that men could not live in society without laws and constraint". Significant iconoclastic riots took place in Basel (in 1529), Zurich (1523), Copenhagen (1530), Münster (1534), Geneva (1535), Augsburg (1537), Scotland (1559), Rouen (1560), and Saintes and La Rochelle (1562). Calvinist iconoclasm in Europe "provoked reactive riots by Lutheran mobs" in Germany and "antagonized the neighbouring Eastern Orthodox" in the Baltic region. The Seventeen Provinces (now the Netherlands, Belgium, and parts of Northern France) were disrupted by widespread Calvinist iconoclasm in the summer of 1566. This period, known as the Beeldenstorm, began with the destruction of the statuary of the Monastery of Saint Lawrence in Steenvoorde after a "Hagenpreek," or field sermon, by Sebastiaan Matte on 10 August 1566; by October the wave of furor had gone all through the Spanish Netherlands up to Groningen. Hundreds of other attacks included the sacking of the Monastery of Saint Anthony after a sermon by Jacob de Buysere. The Beeldenstorm marked the start of the revolution against the Spanish forces and the Catholic Church. During the Reformation in England, which started during the reign of Anglican monarch Henry VIII, and was urged on by reformers such as Hugh Latimer and Thomas Cranmer, limited official action was taken against religious images in churches in the late 1530s. Henry's young son, Edward VI, came to the throne in 1547 and, under Cranmer's guidance, issued injunctions for Religious Reforms in the same year and in 1550, an Act of Parliament "for the abolition and putting away of divers books and images." During the English Civil War, the Parliamentarians reorganised the administration of East Anglia into the Eastern Association of counties. This covered some of the wealthiest counties in England, which in turn financed a substantial and significant military force. After Earl of Manchester was appointed the commanding officer of these forces, and in turn he appointed Smasher Dowsing as Provost Marshal, with a warrant to demolish religious images which were considered to be superstitious or linked with popism. Bishop Joseph Hall of Norwich described the events of 1643 when troops and citizens, encouraged by a Parliamentary ordinance against superstition and idolatry, behaved thus: Lord what work was here! What clattering of glasses! What beating down of walls! What tearing up of monuments! What pulling down of seats! What wresting out of irons and brass from the windows! What defacing of arms! What demolishing of curious stonework! What tooting and piping upon organ pipes! And what a hideous triumph in the market-place before all the country, when all the mangled organ pipes, vestments, both copes and surplices, together with the leaden cross which had newly been sawn down from the Green-yard pulpit and the service-books and singing books that could be carried to the fire in the public market-place were heaped together. Protestant Christianity was not uniformly hostile to the use of religious images. Martin Luther taught the "importance of images as tools for instruction and aids to devotion," stating: "If it is not a sin but good to have the image of Christ in my heart, why should it be a sin to have it in my eyes?" Lutheran churches retained ornate church interiors with a prominent crucifix, reflecting their high view of the real presence of Christ in Eucharist. As such, "Lutheran worship became a complex ritual choreography set in a richly furnished church interior." For Lutherans, "the Reformation renewed rather than removed the religious image." Lutheran scholar Jeremiah Ohl writes:Zwingli and others for the sake of saving the Word rejected all plastic art; Luther, with an equal concern for the Word, but far more conservative, would have all the arts to be the servants of the Gospel. "I am not of the opinion" said [Luther], "that through the Gospel all the arts should be banished and driven away, as some zealots want to make us believe; but I wish to see them all, especially music, in the service of Him Who gave and created them." Again he says: "I have myself heard those who oppose pictures, read from my German Bible.... But this contains many pictures of God, of the angels, of men, and of animals, especially in the Revelation of St. John, in the books of Moses, and in the book of Joshua. We therefore kindly beg these fanatics to permit us also to paint these pictures on the wall that they may be remembered and better understood, inasmuch as they can harm as little on the walls as in books. Would to God that I could persuade those who can afford it to paint the whole Bible on their houses, inside and outside, so that all might see; this would indeed be a Christian work. For I am convinced that it is God's will that we should hear and learn what He has done, especially what Christ suffered. But when I hear these things and meditate upon them, I find it impossible not to picture them in my heart. Whether I want to or not, when I hear, of Christ, a human form hanging upon a cross rises up in my heart: just as I see my natural face reflected when I look into water. Now if it is not sinful for me to have Christ's picture in my heart, why should it be sinful to have it before my eyes?The Ottoman Sultan Suleiman the Magnificent, who had pragmatic reasons to support the Dutch Revolt (the rebels, like himself, were fighting against Spain) also completely approved of their act of "destroying idols," which accorded well with Muslim teachings. A bit later in Dutch history, in 1627 the artist Johannes van der Beeck was arrested and tortured, charged with being a religious non-conformist and a blasphemer, heretic, atheist, and Satanist. The 25 January 1628 judgment from five noted advocates of The Hague pronounced him guilty of "blasphemy against God and avowed atheism, at the same time as leading a frightful and pernicious lifestyle. At the court's order his paintings were burned, and only a few of them survive." Other instances From the 16th through the 19th centuries, many of the polytheistic religious deities and texts of pre-colonial Americas, Oceania, and Africa were destroyed by Christian missionaries and their converts, such as during the Spanish conquest of the Aztec Empire and the Spanish conquest of the Inca Empire. Many of the moai of Easter Island were toppled during the 18th century in the iconoclasm of civil wars before any European encounter. Other instances of iconoclasm may have occurred throughout Eastern Polynesia during its conversion to Christianity in the 19th century. After the Second Vatican Council in the late 20th century, some Roman Catholic parish churches discarded much of their traditional imagery, art, and architecture. Muslim iconoclasm Islam has a much stronger tradition of forbidding the depiction of figures, especially religious figures, with Sunni Islam forbidding it more than Shia Islam. In the history of Islam, the act of removing idols from the Ka'ba in Mecca has great symbolic and historic importance for all believers. In general, Muslim societies have avoided the depiction of living beings (both animals and humans) within such sacred spaces as mosques and madrasahs. This ban on figural representation is not based on the Qur'an, instead, it is based on traditions which are described within the Hadith. The prohibition of figuration has not always been extended to the secular sphere, and a robust tradition of figural representation exists within Muslim art. However, Western authors have tended to perceive "a long, culturally determined, and unchanging tradition of violent iconoclastic acts" within Islamic society. Early Islam in Arabia The first act of Muslim iconoclasm dates to the beginning of Islam, in 630, when the various statues of Arabian deities housed in the Kaaba in Mecca were destroyed. There is a tradition that Muhammad spared a fresco of Mary and Jesus. This act was intended to bring an end to the idolatry which, in the Muslim view, characterized Jahiliyyah. The destruction of the idols of Mecca did not, however, determine the treatment of other religious communities living under Muslim rule after the expansion of the caliphate. Most Christians under Muslim rule, for example, continued to produce icons and to decorate their churches as they wished. A major exception to this pattern of tolerance in early Islamic history was the "Edict of Yazīd", issued by the Umayyad caliph Yazīd II in 722–723. This edict ordered the destruction of crosses and Christian images within the territory of the caliphate. Researchers have discovered evidence that the order was followed, particularly in present-day Jordan, where archaeological evidence shows the removal of images from the mosaic floors of some, although not all, of the churches that stood at this time. But Yazīd's iconoclastic policies were not continued by his successors, and Christian communities of the Levant continued to make icons without significant interruption from the sixth century to the ninth. Egypt Al-Maqrīzī, writing in the 15th century, attributes the missing nose on the Great Sphinx of Giza to iconoclasm by Muhammad Sa'im al-Dahr, a Sufi Muslim in the mid-1300s. He was reportedly outraged by local Muslims making offerings to the Great Sphinx in the hope of controlling the flood cycle, and he was later executed for vandalism. However, whether this was actually the cause of the missing nose has been debated by historians. Mark Lehner, having performed an archaeological study, concluded that it was broken with instruments at an earlier unknown time between the 3rd and 10th centuries. Ottoman conquests Certain conquering Muslim armies have used local temples or houses of worship as mosques. An example is Hagia Sophia in Istanbul (formerly Constantinople), which was converted into a mosque in 1453. Most icons were desecrated and the rest were covered with plaster. In the 1934 the government of Turkey decided to convert the Hagia Sophia into a museum and the restoration of the mosaics was undertaken by the American Byzantine Institute beginning in 1932. Contemporary events Certain Muslim denominations continue to pursue iconoclastic agendas. There has been much controversy within Islam over the recent and apparently on-going destruction of historic sites by Saudi Arabian authorities, prompted by the fear they could become the subject of "idolatry." A recent act of iconoclasm was the 2001 destruction of the giant Buddhas of Bamyan by the then-Taliban government of Afghanistan. The act generated worldwide protests and was not supported by other Muslim governments and organizations. It was widely perceived in the Western media as a result of the Muslim prohibition against figural decoration. Such an account overlooks "the coexistence between the Buddhas and the Muslim population that marveled at them for over a millennium" before their destruction. According to art historian F. B. Flood, analysis of the Taliban's statements regarding the Buddhas suggest that their destruction was motivated more by political than by theological concerns. Taliban spokespeople have given many different explanations of the motives for the destruction. During the Tuareg rebellion of 2012, the radical Islamist militia Ansar Dine destroyed various Sufi shrines from the 15th and 16th centuries in the city of Timbuktu, Mali. In 2016, the International Criminal Court (ICC) sentenced Ahmad al-Faqi al-Mahdi, a former member of Ansar Dine, to nine years in prison for this destruction of cultural world heritage. This was the first time that the ICC convicted a person for such a crime. The short-lived Islamic State of Iraq and the Levant carried out iconoclastic attacks such as the destruction of Shia mosques and shrines. Notable incidents include blowing up the Mosque of the Prophet Yunus (Jonah) and destroying the Shrine to Seth in Mosul. Iconoclasm in India In early Medieval India, there were numerous recorded instances of temple desecration by Indian kings against rival Indian kingdoms, which involved conflicts between devotees of different Hindu deities, as well as conflicts between Hindus, Buddhists, and Jains. In 642, the Pallava king Narasimhavarman I looted a Ganesha temple in the Chalukyan capital of Vatapi. In c. 692, Chalukya armies invaded northern India where they looted temples of Ganga and Yamuna. In the 8th century, Bengali troops from the Buddhist Pala Empire desecrated temples of Vishnu, the state deity of Lalitaditya's kingdom in Kashmir. In the early 9th century, Indian Hindu kings from Kanchipuram and the Pandyan king Srimara Srivallabha looted Buddhist temples in Sri Lanka. In the early 10th century, the Pratihara king Herambapala looted an image from a temple in the Sahi kingdom of Kangra, which was later looted by the Pratihara king Yashovarman. During the Muslim conquest of Sindh Records from the campaign recorded in the Chach Nama record the destruction of temples during the early 8th century when the Umayyad governor of Damascus, al-Hajjaj ibn Yusuf, mobilized an expedition of 6000 cavalry under Muhammad bin Qasim in 712. Historian Upendra Thakur records the persecution of Hindus and Buddhists: The Somnath temple and Mahmud of Ghazni Perhaps the most notorious episode of iconoclasm in India was Mahmud of Ghazni's attack on the Somnath Temple from across the Thar Desert. The temple was first raided in 725, when Junayad, the governor of Sind, sent his armies to destroy it. In 1024, during the reign of Bhima I, the prominent Turkic-Muslim ruler Mahmud of Ghazni raided Gujarat, plundering the Somnath Temple and breaking its jyotirlinga despite pleas by Brahmins not to break it. He took away a booty of 20 million dinars. The attack may have been inspired by the belief that an idol of the goddess Manat had been secretly transferred to the temple. According to the Ghaznavid court-poet Farrukhi Sistani, who claimed to have accompanied Mahmud on his raid, Somnat (as rendered in Persian) was a garbled version of su-manat referring to the goddess Manat. According to him, as well as a later Ghaznavid historian Abu Sa'id Gardezi, the images of the other goddesses were destroyed in Arabia but the one of Manat was secretly sent away to Kathiawar (in modern Gujarat) for safekeeping. Since the idol of Manat was an aniconic image of black stone, it could have been easily confused with a lingam at Somnath. Mahmud is said to have broken the idol and taken away parts of it as loot and placed so that people would walk on it. In his letters to the Caliphate, Mahmud exaggerated the size, wealth and religious significance of the Somnath temple, receiving grandiose titles from the Caliph in return. The wooden structure was replaced by Kumarapala (r. 1143–72), who rebuilt the temple out of stone. From the Mamluk dynasty onward Historical records which were compiled by the Muslim historian Maulana Hakim Saiyid Abdul Hai attest to the religious violence which occurred during the Mamluk dynasty under Qutb-ud-din Aybak. The first mosque built in Delhi, the "Quwwat al-Islam" was built with demolished parts of 20 Hindu and Jain temples. This pattern of iconoclasm was common during his reign. During the Delhi Sultanate, a Muslim army led by Malik Kafur, a general of Alauddin Khalji, pursued four violent campaigns into south India, between 1309 and 1311, against the Hindu kingdoms of Devgiri (Maharashtra), Warangal (Telangana), Dwarasamudra (Karnataka) and Madurai (Tamil Nadu). Many Temples were plundered; Hoysaleswara Temple and others were ruthlessly destroyed. In Kashmir, Sikandar Shah Miri (1389–1413) began expanding, and unleashed religious violence that earned him the name but-shikan, or 'idol-breaker'. He earned this sobriquet because of the sheer scale of desecration and destruction of Hindu and Buddhist temples, shrines, ashrams, hermitages, and other holy places in what is now known as Kashmir and its neighboring territories. Firishta states, "After the emigration of the Brahmins, Sikundur ordered all the temples in Kashmeer to be thrown down." He destroyed vast majority of Hindu and Buddhist temples in his reach in Kashmir region (north and northwest India). In the 1460s, Kapilendra, founder of the Suryavamsi Gajapati dynasty, sacked the Shaiva and Vaishnava temples in the Cauvery delta in the course of wars of conquest in the Tamil country. Vijayanagara king Krishnadevaraya looted a Bala Krishna temple in Udayagiri in 1514, and looted a Vitthala temple in Pandharpur in 1520. A regional tradition, along with the Hindu text Madala Panji, states that Kalapahar attacked and damaged the Konark Sun Temple in 1568, as well as many others in Orissa. Some of the most dramatic cases of iconoclasm by Muslims are found in parts of India where Hindu and Buddhist temples were razed and mosques erected in their place. Aurangzeb, the 6th Mughal Emperor, destroyed the famous Hindu temples at Varanasi and Mathura, turning back on his ancestor Akbar's policy of religious freedom and establishing Sharia across his empire. During the Goa Inquisition Exact data on the nature and number of Hindu temples destroyed by the Christian missionaries and Portuguese government are unavailable. Some 160 temples were allegedly razed to the ground in Tiswadi (Ilhas de Goa) by 1566. Between 1566 and 1567, a campaign by Franciscan missionaries destroyed another 300 Hindu temples in Bardez (North Goa). In Salcete (South Goa), approximately another 300 Hindu temples were destroyed by the Christian officials of the Inquisition. Numerous Hindu temples were destroyed elsewhere at Assolna and Cuncolim by Portuguese authorities. A 1569 royal letter in Portuguese archives records that all Hindu temples in its colonies in India had been burnt and razed to the ground. The English traveller Sir Thomas Herbert, 1st Baronet who visited Goa in the 1600s writes: In modern India 1) Dr. Ambedkar and his supporters on 25 December 1927 in the Mahad Satyagraha strongly criticised, condemned and then burned copies of Manusmriti on a pyre in a specially dug pit. Manusmriti, one of the sacred Hindu texts, is the religious basis of casteist laws and values of Hinduism and hence was/is the reason of social and economic plight of crores of untouchables and lower caste Hindus. One of the greatest iconoclasts for all time, this explosive incident rocked the Hindu society. Ambedkarites continue to observe 25 December as "Manusmriti Dahan Divas" (Manusmriti Burning Day) and burn copies of Manusmriti on this day. 2) The most high-profile case of Independent India was in 1992. Hindus, led by the Vishva Hindu Parishad and Bajrang Dal, destroyed the 430-year-old Islamic Babri Masjid in Ayodhya which was built after destroying the Ram Mandir. This was to reclaim their temple which was destroyed by Islamic iconoclasts. Iconoclasm in East Asia China There have been a number of anti-Buddhist campaigns in Chinese history that led to the destruction of Buddhist temples and images. One of the most notable of these campaigns was the Great Anti-Buddhist Persecution of the Tang dynasty. During and after the 1911 Xinhai Revolution, there was widespread destruction of religious and secular images in China. During the Northern Expedition in Guangxi in 1926, Kuomintang General Bai Chongxi led his troops in destroying Buddhist temples and smashing Buddhist images, turning the temples into schools and Kuomintang party headquarters. It was reported that almost all of the viharas in Guangxi were destroyed and the monks were removed. Bai also led a wave of anti-foreignism in Guangxi, attacking Americans, Europeans, and other foreigners, and generally making the province unsafe for foreigners and missionaries. Westerners fled from the province and some Chinese Christians were also attacked as imperialist agents. The three goals of the movement were anti-foreignism, anti-imperialism and anti-religion. Bai led the anti-religious movement against superstition. Huang Shaohong, also a Kuomintang member of the New Guangxi clique, supported Bai's campaign. The anti-religious campaign was agreed upon by all Guangxi Kuomintang members. There was extensive destruction of religious and secular imagery in Tibet after it was invaded and occupied by China. Many religious and secular images were destroyed during the Cultural Revolution of 1966–1976, ostensibly because they were a holdover from China's traditional past (which the Communist regime led by Mao Zedong reviled). The Cultural Revolution included widespread destruction of historic artworks in public places and private collections, whether religious or secular. Objects in state museums were mostly left intact. South Korea According to an article in Buddhist-Christian Studies:Over the course of the last decade [1990s] a fairly large number of Buddhist temples in South Korea have been destroyed or damaged by fire by Christian fundamentalists. More recently, Buddhist statues have been identified as idols, and attacked and decapitated in the name of Jesus. Arrests are hard to effect, as the arsonists and vandals work by stealth of night. Angkor Beginning around 1243 AD with the death of Indravarman II, the Khmer Empire went through a period of iconoclasm. At the beginning of the reign of the next king, Jayavarman VIII, the Kingdom went back to Hinduism and the worship of Shiva. Many of the Buddhist images were destroyed by Jayavarman VIII, who reestablished previously Hindu shrines that had been converted to Buddhism by his predecessor. Carvings of the Buddha at temples such as Preah Khan were destroyed, and during this period the Bayon Temple was made a temple to Shiva, with the central 3.6 meter tall statue of the Buddha cast to the bottom of a nearby well. Political iconoclasm Damnatio memoriae Revolutions and changes of regime, whether through uprising of the local population, foreign invasion, or a combination of both, are often accompanied by the public destruction of statues and monuments identified with the previous regime. This may also be known as damnatio memoriae, the ancient Roman practice of official obliteration of the memory of a specific individual. Stricter definitions of "iconoclasm" exclude both types of action, reserving the term for religious or more widely cultural destruction. In many cases, such as Revolutionary Russia or Ancient Egypt, this distinction can be hard to make. Among Roman emperors and other political figures subject to decrees of damnatio memoriae were Sejanus, Publius Septimius Geta, and Domitian. Several Emperors, such as Domitian and Commodus had during their reigns erected numerous statues of themselves, which were pulled down and destroyed when they were overthrown. The perception of damnatio memoriae in the Classical world was an act of erasing memory has been challenged by scholars who have argued that it "did not negate historical traces, but created gestures which served to dishonor the record of the person and so, in an oblique way, to confirm memory," and was in effect a spectacular display of "pantomime forgetfulness." Examining cases of political monument destruction in modern Irish history, Guy Beiner has demonstrated that iconoclastic vandalism often entails subtle expressions of ambiguous remembrance and that, rather than effacing memory, such acts of de-commemorating effectively preserve memory in obscure forms. During the French Revolution Throughout the radical phase of the French Revolution, iconoclasm was supported by members of the government as well as the citizenry. Numerous monuments, religious works, and other historically significant pieces were destroyed in an attempt to eradicate any memory of the Old Regime. A statue of King Louis XV in the Paris square which until then bore his name, was pulled down and destroyed. This was a prelude to the guillotining of his successor Louis XVI in the same site, renamed "Place de la Révolution" (at present Place de la Concorde). Later that year, the bodies of many French kings were exhumed from the Basilica of Saint-Denis and dumped in a mass grave. Some episodes of iconoclasm were carried out spontaneously by crowds of citizens, including the destruction of statues of kings during the insurrection of 10 August 1792 in Paris. Some were directly sanctioned by the Republican government, including the Saint-Denis exhumations. Nonetheless, the Republican government also took steps to preserve historic artworks, notably by founding the Louvre museum to house and display the former royal art collection. This allowed the physical objects and national heritage to be preserved while stripping them of their association with the monarchy. Alexandre Lenoir saved many royal monuments by diverting them to preservation in a museum. The statue of Napoleon on the column at Place Vendôme, Paris was also the target of iconoclasm several times: destroyed after the Bourbon Restoration, restored by Louis-Philippe, destroyed during the Paris Commune and restored by Adolphe Thiers. Other examples Other examples of political destruction of images include: There have been several cases of removing symbols of past rulers in Malta's history. Many Hospitaller coats of arms on buildings were defaced during the French occupation of Malta in 1798–1800; a few of these were subsequently replaced by British coats of arms in the early 19th century. Some British symbols were also removed by the government after Malta became a republic in 1974. These include royal cyphers being ground off from post boxes, and British coats of arms such as that on the Main Guard building being temporarily obscured (but not destroyed). With the entry of the Ottoman Empire to the First World War, the Ottoman Army destroyed the Russian victory monument erected in San Stefano (the modern Yeşilköy quarter of Istanbul) to commemorate the Russian victory in the Russo-Turkish War of 1877–1878. The demolition was filmed by former army officer Fuat Uzkınay, producing Ayastefanos'taki Rus Abidesinin Yıkılışı—the oldest known Turkish-made film. In the late 18th century, the Brabant Revolutionaries sacked Brussels' Grand-Place, destroying statues of nobility and symbols of Christianity. In the 19th century, the place was renovated and many new statues added. In 1911, a marble commemoration for the Spanish freethinker and educator Francisco Ferrer, executed two years earlier and widely considered a martyr, was erected in the Grand-Place. The statue depicted a nude man holding the Torch of Enlightenment. The Imperial German military, which occupied Belgium during the First World War, disliked the monument and destroyed it in 1915. It was restored in 1926 by the International Free Thought Movement. In 1942 the pro-Nazi Vichy Government of France took down and melted Clothilde Roch's statue of the 16th-century dissident intellectual Michael Servetus, who had been burned at the stake in Geneva at the instigation of Calvin. The Vichy authorities disliked the statue, as it was a celebration of freedom of conscience. In 1960, having found the original molds, the municipality of Annemasse had it recast and returned the statue to its previous place. A sculpture of the head of Spanish intellectual Miguel de Unamuno by Victorio Macho was installed in the City Hall of Bilbao, Spain. It was withdrawn in 1936 when Unamuno showed temporary support for the Nationalist side. During the Spanish Civil War, it was thrown into the estuary. It was later recovered. In 1984 the head was installed in Plaza Unamuno. In 1999, it was again thrown into the estuary after a political meeting of . It was substituted by a copy in 2000 after the original was located in the water. The Battle of Baghdad and the regime of Saddam Hussein symbolically ended with the Firdos Square statue destruction, a U.S. military-staged event on April 9, 2003 where a prominent statue of Saddam Hussein was pulled down. Subsequently, statues and murals of Saddam Hussein all over Iraq were destroyed by US occupation forces as well as Iraqi citizens. In 2016, paintings from the University of Cape Town were burned in student protests as symbols of colonialism. In November 2019 the statue of Swedish footballer Zlatan Ibrahimović was vandalized by Malmö FF supporters after he announced he had become part-owner of Swedish rivals Hammarby. White paint was sprayed on it; threats and hateful messages towards Zlatan were written on the statue, and it was burned. In a second attack the nose was sawed off and the statue was sprinkled with chrome paint. On 5 January 2020 it was finally toppled. On June 4, 2020, Virginia governor Ralph Northam ordered the monument to Robert E. Lee in Richmond to be removed in response to George Floyd protests. It was removed on September 8, 2021. In the Soviet Union During and after the October Revolution, widespread destruction of religious and secular imagery in Russia took place, as well as the destruction of imagery related to the Imperial family. The Revolution was accompanied by destruction of monuments of tsars, as well as the destruction of imperial eagles at various locations throughout Russia. According to Christopher Wharton:In front of a Moscow cathedral, crowds cheered as the enormous statue of Tsar Alexander III was bound with ropes and gradually beaten to the ground. After a considerable amount of time, the statue was decapitated and its remaining parts were broken into rubble.The Soviet Union actively destroyed religious sites, including Russian Orthodox churches and Jewish cemeteries, in order to discourage religious practice and curb the activities of religious groups. During the Hungarian Revolution of 1956 and during the Revolutions of 1989, protesters often attacked and took down sculptures and images of Joseph Stalin, such as the Stalin Monument in Budapest. The fall of Communism in 1989-1991 was also followed by the destruction or removal of statues of Vladimir Lenin and other Communist leaders in the former Soviet Union and in other Eastern Bloc countries. Particularly well-known was the destruction of "Iron Felix", the statue of Felix Dzerzhinsky outside the KGB's headquarters. Another statue of Dzerzhinsky was destroyed in a Warsaw square that was named after him during communist rule, but which is now called Bank Square. In the United States During the American Revolution, the Sons of Liberty pulled down and destroyed the gilded lead statue of George III of the United Kingdom on Bowling Green (New York City), melting it down to be recast as ammunition. Similar acts have accompanied the independence of most ex-colonial territories. Sometimes relatively intact monuments are moved to a collected display in a less prominent place, as in India and also post-Communist countries. In August 2017, a statue of a Confederate soldier dedicated to "the boys who wore the gray" was pulled down from its pedestal in front of Durham County Courthouse in North Carolina by protesters. This followed the events at the 2017 Unite the Right rally in response to growing calls to remove Confederate monuments and memorials across the U.S. 2020 demonstrations During the George Floyd protests of 2020, demonstrators pulled down dozens of statues which they considered symbols of the Confederacy, slavery, segregation, or racism, including the statue of Williams Carter Wickham in Richmond, Virginia Further demonstrations in the wake of the George Floyd protests have resulted in the removal of: the John Breckenridge Castleman monument in Louisville, Kentucky; plaques in Jacksonville, Florida's Hemming Park (renamed in 1899 in honor of Civil War veteran Charles C. Hemming), which were in remembrance of deceased Confederate soldiers; the monumental obelisk of the Confederate Soldiers and Sailors Monument and a statue of Charles Linn in Linn Park, Birmingham, Alabama; a statue of Junípero Serra in Golden Gate Park, San Francisco; a statue of Confederate Gen. Robert E. Lee in Montgomery, Alabama; the Appomattox statue in Alexandria, Virginia, leaving the monument's base empty but intact. Multiple statues of early European explorers and founders were also vandalized, including those of Christopher Columbus, George Washington, and Thomas Jefferson. Christopher Columbus was removed in Virginia, Minnesota, Chicago and beheaded in Boston MA. George Washington statue was toppled in Portland, Oregon. A statue of the African-American abolitionist statesman Frederick Douglass was vandalised in Rochester, New York, by being torn from its base and left close to a nearby river gorge. Donald Trump attributed the act to anarchists, but he did not substantiate his claim nor did he offer a theory on motive. Cornell William Brooks, former president of the NAACP, theorised that this was an act of revenge from white supremacists. Carvin Eison, who led the project that brought the Douglass statues to Rochester, thought it was unlikely that the Douglass statue was toppled by someone who was upset about monuments honoring Confederate figures, and added that "it's only logical that it was some kind of retaliation event in someone's mind". Police did not find evidence that supported or refuted either claim, and the vandalism case remains unsolved. See also Aniconism Censorship by religion Cult image Cultural Revolution Icon Iconolatry List of destroyed heritage Lost artworks Natural theology Slighting Council of Constantinople (843) Notes References Further reading (On the conceptual background of Byzantine iconoclasm) —— 2016. Broken Idols of the English Reformation. Cambridge University Press. Barasch, Moshe. 1992. Icon: Studies in the History of an Idea. New York University Press. . Besançon, Alain. 2000. The Forbidden Image: An Intellectual History of Iconoclasm. University of Chicago Press. . Bevan, Robert. 2006. The Destruction of Memory: Architecture at War. Reaktion Books. . Boldrick, Stacy, Leslie Brubaker, and Richard Clay, eds. 2014. Striking Images, Iconoclasms Past and Present. Ashgate. (Scholarly studies of the destruction of images from prehistory to the Taliban.) Calisi, Antonio. 2017. I Difensori Dell'icona: La Partecipazione Dei Vescovi Dell'Italia Meridionale Al Concilio Di Nicea II 787. CreateSpace. . Freedberg, David. 1977. "The Structure of Byzantine and European Iconoclasm." Pp. 165–77 in Iconoclasm: Papers Given at the Ninth Spring Symposium of Byzantine Studies, edited by A. Bryer and J. Herrin. University of Birmingham, Centre for Byzantine Studies. . —— [1985] 1993. "Iconoclasts and their Motives," (Second Horst Gerson Memorial Lecture, University of Groningen). Public 8(Fall). Original print: Maarssen: Gary Schwartz. 1985. . Shourie, Arun, Sita Ram Goel, Harsh Narain, Jay Dubashi, and Ram Swarup. 1990. Hindu Temples - What Happened to Them Vol. I, (A Preliminary Survey). Topper, David R. Idolatry & Infinity: Of Art, Math & God. BrownWalker. . Weeraratna, Senaka ' Repression of Buddhism in Sri Lanka by the Portuguese' (1505 -1658) Teodoro Studita, Contro gli avversari delle icone, Emanuela Fogliadini (Prefazione), Antonio Calisi (Traduttore), Jaca Book, 2022, External links Iconoclasm in England, Holy Cross College (UK) Design as Social Agent at the ICA by Kerry Skemp, April 5, 2009 Hindu temples destroyed by Muslim rulers in India Protestant Reformation Religious persecution Christian terminology
15260
https://en.wikipedia.org/wiki/Islands%20of%20the%20North%20Atlantic
Islands of the North Atlantic
IONA (Islands of the North Atlantic) is an acronym suggested in 1980 by Sir John Biggs-Davison to refer to a loose linkage of the Channel Islands (Guernsey and Jersey), Great Britain (England, Scotland, and Wales), Ireland (Northern Ireland and the Republic of Ireland), and the Isle of Man, similar to the present day British–Irish Council. Its intended purpose was as a more politically acceptable alternative to the British Isles, which is disliked by some people in Ireland. The neologism has been criticised on the grounds that it excludes most of the islands in the North Atlantic, and also that the only island referred to by the term that is actually in the North Atlantic Ocean is Ireland (Great Britain is in fact in between the Irish Sea and The North Sea.) In the context of the Northern Irish peace process, during the negotiation of the Good Friday Agreement, IONA was unsuccessfully proposed as a neutral name for the proposed council. One feature of this name is that IONA has the same spelling as the island of Iona which is off the coast of Scotland, but with which Irish people have strong cultural associations. It is therefore a name with which people of both main islands might identify. Taoiseach Bertie Ahern noted the symbolism in a 2006 address in Edinburgh:[The Island of] Iona is a powerful symbol of relationships between these islands, with its ethos of service not dominion. Iona also radiated out towards the Europe of the Dark Ages, not to mention Pagan England at Lindisfarne. The British-Irish Council is the expression of a relationship that at the origin of the Anglo-Irish process in 1981 was sometimes given the name Iona, islands of the North Atlantic, and sometimes Council of the Isles, with its evocation of the Lords of the Isles of the 14th and 15th centuries who spanned the North Channel. In the 17th century, Highland warriors and persecuted Presbyterian Ministers criss-crossed the North Channel. In a Dáil Éireann debate, Proinsias De Rossa was less enthusiastic: The acronym IONA is a useful way of addressing the coming together of these two islands. However, the island of Iona is probably a green heaven in that nobody lives on it and therefore it cannot be polluted in any way. The term IONA is used by the World Universities Debating Championship. IONA is one of the regions that appoint a representative onto the committee of the World Universities Debating Council. Greenland, the Faroe Islands and Iceland are included in the definition of IONA used in this context, while Newfoundland and Prince Edward Island are in the North American region. However, none of these islands have yet participated in the World Universities Debating Championships. Otherwise, the term has achieved very little popular usage in any context. See also British Isles naming dispute Terminology of the British Isles References External links The Good Friday Agreement British-Irish Council Project IONA Terminology of the British Isles Islands of the North Atlantic Ocean Geographical neologisms
15437
https://en.wikipedia.org/wiki/ITU%20prefix
ITU prefix
The International Telecommunication Union (ITU) allocates call sign prefixes for radio and television stations of all types. They also form the basis for, but may not exactly match, aircraft registration identifiers. These prefixes are agreed upon internationally, and are a form of country code. A call sign can be any number of letters and numerals but each country must only use call signs that begin with the characters allocated for use in that country. With regard to the second and/or third letters in the prefixes in the list below, if the country in question is allocated all callsigns with A to Z in that position, then that country can also use call signs with the digits 0 to 9 in that position. For example, the United States is assigned KA–KZ, and therefore can also use prefixes like K1 or K9. While ITU prefix rules are adhered to in the context of international broadcasting, including amateur radio, it is rarer for countries to assign broadcast call signs to conventional AM, FM, and television stations with purely domestic reach; the United States, Canada, Mexico, Japan, South Korea, and the Philippines are among those that do. Canada presents one notable exception to the ITU prefix rules: it uses CB for its own Canadian Broadcasting Corporation stations, whereas Chile is officially assigned the CB prefix. Innovation, Science and Economic Development Canada's broadcasting rules indicate this is through a "special arrangement", without elaborating. In any case, the two countries are geographically separate enough to prevent confusion; Canada's shortwave broadcasters and amateur radio stations have always used one of its assigned ITU prefixes. Unallocated and unavailable call sign prefixes Unallocated: The following call sign prefixes are available for future allocation by the ITU. (x represents any letter; n represents any digit from 2–9.) E8, E9, H5, J9, On, S4, T9*, Un, V9, Xn, YZ*, Z4–Z7, Z9, 4N*. (* Indicates a prefix that has recently been returned to the ITU.) Unavailable: Under present ITU guidelines the following call sign prefixes shall not be allocated. They are sometimes used unofficially – such as amateur radio operators operating in a disputed territory or in a nation state that has no official prefix (e.g. S0 in Western Sahara, station 1A0 at Knights of Malta headquarters in Rome, or station 1L in Liberland). (x represents any letter; n represents any digit from 2–9.) nn, x0, x1, 0x, 1x, Qx. no prefixes beginning with Q are used—they may be confused with Q codes. Note that this applies to prefixes only - suffixes are the responsibility of the allocating country. no prefixes with the digits 1 or 0 are used—they may be confused with the letters I or O. two digit prefixes (nn) are not as yet considered by the ITU. Allocation table See also Call-sign allocation plan Non-ITU prefix Amateur radio call signs Aircraft registration Notes References External links Call signs Amateur radio call signs
15642
https://en.wikipedia.org/wiki/January
January
January is the first month of the year in the Julian and Gregorian calendars and is also the first of seven months to have a length of 31 days. The first day of the month is known as New Year's Day. It is, on average, the coldest month of the year within most of the Northern Hemisphere (where it is the second month of winter) and the warmest month of the year within most of the Southern Hemisphere (where it is the second month of summer). In the Southern hemisphere, January is the seasonal equivalent of July in the Northern hemisphere and vice versa. Ancient Roman observances during this month include Cervula and Juvenalia, celebrated January 1, as well as one of three Agonalia, celebrated January 9, and Carmentalia, celebrated January 11. These dates do not correspond to the modern Gregorian calendar. History January (in Latin, Ianuarius) is named after Janus, the god of beginnings and transitions in Roman mythology. Traditionally, the original Roman calendar consisted of 10 months totaling 304 days, winter being considered a month-less period. Around 713 BC, the semi-mythical successor of Romulus, King Numa Pompilius, is supposed to have added the months of January and February, so that the calendar covered a standard lunar year (354 days). Although March was originally the first month in the old Roman calendar, January became the first month of the calendar year either under Numa or under the Decemvirs about 450 BC (Roman writers differ). In contrast, each specific calendar year was identified by the names of the two consuls, who entered office on March 15 until 153 BC, at which point they started entering office on January 1. Various Christian feast dates were used for the New Year in Europe during the Middle Ages, including March 25 (Feast of the Annunciation) and December 25. However, medieval calendars were still displayed in the Roman fashion with twelve columns from January to December. Beginning in the 16th century, European countries began officially making January 1 the start of the New Year once again—sometimes called Circumcision Style because this was the date of the Feast of the Circumcision, being the seventh day after December 25. Historical names for January include its original Roman designation, Ianuarius, the Saxon term Wulf-monath (meaning "wolf month") and Charlemagne's designation Wintarmanoth ("winter / cold month"). In Slovene, it is traditionally called prosinec. The name, associated with millet bread and the act of asking for something, was first written in 1466 in the Škofja Loka manuscript. According to Theodor Mommsen, 1 January became the first day of the year in 600 AUC of the Roman calendar (153 BC), due to disasters in the Lusitanian War. A Lusitanian chief called Punicus invaded the Roman territory, defeated two Roman governors, and killed their troops. The Romans resolved to send a consul to Hispania, and in order to accelerate the dispatch of aid, "they even made the new consuls enter into office two months and a half before the legal time" (March 15). January symbols January's birthstone is the garnet, which represents constancy. Its birth flower is the cottage pink Dianthus caryophyllus or galanthus as well as the traditional carnation. The Japanese floral emblem of January is the camellia (Camellia sinensis). In Finnish, the month of tammikuu means the heart of the winter and because the name literally means "oak moon", it can be inferred that the oak tree is the heart of the grand forest with many valuable trees as opposed to the typical Arctic forests, which are typically pine and spruce. The photograph of a large tree covered with ice against a blue sky is a familiar scene during Finland's winter. The zodiac signs for the month of January are Capricorn (until January 19) and Aquarius (January 20 onwards). The name of the full moon occurring in January is the wolf moon. January observances This list does not necessarily imply either official status or general observance. Month-long observances Alzheimer's Awareness Month (Canada) Dry January (United Kingdom) National Codependency Awareness Month (United States) National Mentoring Month (United States) National Healthy Weight Awareness Month (United States) Slavery and Human Trafficking Prevention Month (United States) Stalking Awareness Month (United States) Veganuary Food months in the United States This list does not necessarily imply either official status or general observance. Be Kind to Food Servers Month (by proclamation, State of Tennessee) California Dried Plum Digestive Health Month Hot Tea Month National Soup Month Oatmeal Month Non-Gregorian observances All Baha'i, Islamic, and Jewish observances begin at sundown prior to the date listed, and end at sundown on the date in question. List of observances set by the Bahá'í calendar List of observances set by the Chinese calendar List of observances set by the Hebrew calendar List of observances set by the Islamic calendar List of observances set by the Solar Hijri calendar Moveable observances This list does not necessarily imply either official status or general observance. See: List of movable Western Christian observances See: List of movable Eastern Christian observances January 2 unless that day is a Sunday, in which case January 3 New Year Holiday (Scotland) First Friday Children's Day (Bahamas) Second Saturday Children's Day (Thailand) Second Monday Birthday of Eugenio María de Hostos (Puerto Rico, United States) Coming of Age Day (Japan) Friday before third Monday Lee–Jackson Day (Virginia, United States, defunct) Third Friday International Fetish Day Sunday closest to January 22 National Sanctity of Human Life Day (United States) Third full week of January Hunt for Happiness Week (International observance) National Non-Smoking Week (Canada) Last full week of January National School Choice Week (United States) Third Monday Martin Luther King, Jr. Day (United States) Idaho Human Rights Day (Idaho, United States) Wednesday of the third full week of January Weedless Wednesday (Canada) Friday between January 19–25 Husband's Day (Iceland) Last Saturday National Seed Swap Day (United States) Last Sunday Liberation of Auschwitz Memorial Day (Netherlands) January 30 or the nearest Sunday World Leprosy Day Last Monday in January Bubble Wrap Appreciation Day Fourth Monday Community Manager Appreciation Day (International observance) National Heroes' Day (Cayman Islands) Monday Closest to January 29 Auckland Anniversary Day Fixed observances December 25 – January 5: Twelve Days of Christmas (Western Christianity) December 26 – January 1: Kwanzaa (African Americans) December 31 – January 1, in some cases until January 2: Hogmanay (Scotland) January 1 Feast of the Circumcision of Christ Feast of the Holy Name of Jesus (Anglican Communion, Lutheran Church) Feast of Fools (Medieval Europe) Constitution Day (Italy) Dissolution of Czechoslovakia-related observances: Day of the Establishment of the Slovak Republic (Slovakia) Restoration Day of the Independent Czech State (Czech Republic) Euro Day (European Union) Flag Day (Lithuania) Founding Day (Taiwan) Global Family Day Independence Day (Brunei, Cameroon, Haiti, Sudan) International Nepali Dhoti and Nepali Topi Day Jump-up Day (Montserrat, British Overseas Territories) Kalpataru Day (Ramakrishna Movement) National Bloody Mary Day (United States) National Tree Planting Day (Tanzania) New Year's Day Japanese New Year Novy God Day (Russia) Sjoogwachi (Okinawa Islands) Polar Bear Swim Day (Canada and United States) Public Domain Day (multiple countries) Solemnity of Mary, Mother of God (Catholic Church) World Day of Peace Triumph of the Revolution (Cuba) January 2 Ancestry Day (Haiti) Berchtold's Day (Liechtenstein, Switzerland, and the Alsace) Carnival Day (Saint Kitts and Nevis) Happy Mew Year For Cats Day (Thomas Roy) Kakizome (Japan) National Creampuff Day (United States) National Science Fiction Day (United States) The second day of New Year (a holiday in Armenia, Kazakhstan, North Macedonia, Mauritius, Montenegro, New Zealand, Romania, Russia, Switzerland, and Ukraine) Nyinlong (Bhutan) Victory of Armed Forces Day (Cuba) January 3 Anniversary of the 1966 Coup d’état (Burkina Faso) Memento Mori (Thomas Roy) Nakhatsenendyan toner (Armenia): January 3–5 Ministry of Religious Affairs Day (Indonesia) National Chocolate Covered Cherry Day (United States) Tamaseseri Festival (Hakozaki Shrine, Fukuoka, Japan) January 4 Day of the Fallen against the Colonial Repression (Angola) Day of the Martyrs (Democratic Republic of the Congo) Dimpled Chad Day (Thomas Roy) Hwinukan mukee (Okinawa Islands, Japan) Independence Day (Myanmar) Ogoni Day (Movement for the Survival of the Ogoni People) World Braille Day January 5 National Bird Day (United States) National Whipped Cream Day (United States) Sausage Day (United Kingdom) Strawberry day (Japan) Take Our Daughters and Sons to Work Day (Sydney, Melbourne, and Brisbane, Australia) Tucindan (Serbia, Montenegro) January 6 Armed Forces Day (Iraq) Epiphany or Three Kings' Day (Western Christianity) or Theophany (Eastern Christianity), and its related observances: Befana Day (Italy) Christmas (Armenian Apostolic Church) Christmas Eve (Russia) Christmas Eve (Ukraine) Christmas Eve (Bosnia and Herzegovina) Christmas Eve (North Macedonia) Little Christmas (Ireland) Þrettándinn (Iceland) Three Wise Men Day Pathet Lao Day (Laos) January 7 Christmas (Eastern Orthodox Churches and Oriental Orthodox Churches using the Julian Calendar, Rastafari) Christmas in Russia Christmas in Ukraine Christmas (Bosnia and Herzegovina) Remembrance Day of the Dead (Armenia) Distaff Day (Medieval Europe) Nanakusa no sekku (Japan) Pioneer's Day (Liberia) Tricolour day (Italy) Victory from Genocide Day (Cambodia) January 8 Show and Tell Day at Work (Thomas Roy) The Eighth (United States) (defunct observance) Typing Day (international observance) January 9 Start of Hōonkō (Nishi Honganji) January 9–16 (Jōdo Shinshū Buddhism) Martyrs' Day (Panama) National Cassoulet Day (United States) Non-Resident Indian Day (India) Republic Day (Republika Srpska) (defunct, declared unconstitutional by the Constitutional Court of Bosnia and Herzegovina) St. Stephen's Day (Eastern Orthodox) January 10 Fête du Vodoun (Benin) Majority Rule Day (Bahamas) January 11 Children's Day (Tunisia) Eugenio María de Hostos Day (Puerto Rico) German Apples Day (Germany) Independence Manifesto Day (Morocco) Kagami biraki (Japan) National Human Trafficking Awareness Day (United States) Republic Day (Albania) January 12 Memorial Day (Turkmenistan) Prosecutor General's Day (Russia) National Youth Day (India) Zanzibar Revolution Day (Tanzania) January 13 Constitution Day (Mongolia) Democracy Day (Cape Verde) Liberation Day (Togo) Old New Year's Eve (Russia, Belarus, Ukraine, Serbia, Montenegro, Republic of Srpska, North Macedonia), and its related observances: Malanka (Ukraine, Russia, Belarus) St. Knut's Day (Sweden and Finland) Stephen Foster Memorial Day (United States) January 14 Azhyrnykhua (Abkhazia) Day of Defenders of the Motherland (Uzbekistan) Feast of Divina Pastora (Barquisimeto) Feast of the Ass (Medieval Christianity) Flag Day (Georgia) National Forest Conservation Day (Thailand) Ratification Day (United States) Revolution and Youth Day (Tunisia) Yennayer (Berbers) January 15 Arbor Day (Egypt) Armed Forces Day (Nigeria) Indian Army Day (India) John Chilembwe Day (Malawi) Korean Alphabet Day (North Korea) Sagichō at Tsurugaoka Hachimangū (Kamakura, Japan) Teacher's Day (Venezuela) Wikipedia Day (international observance) January 16 National Nothing Day (Thomas Roy) National Religious Freedom Day (United States) Solemnity of Mary, Mother of God (Coptic Church) Teacher's Day (Myanmar) Teachers' Day (Thailand) Zuuruku Nichi (Okinawa Islands, Japan) Thiruvalluvar Day (Tamil Nadu, India) January 17 Hardware Freedom Day (international observance) Judgement Day National Day (Menorca) The opening ceremony of Patras Carnival, celebrated until Clean Monday (Patras) January 18 Revolution and Youth Day (Tunisia) Royal Thai Armed Forces Day (Thailand) Week of Prayer for Christian Unity (January 18–25) (Christianity) January 19 Confederate Heroes Day (Texas), and its related observance: Robert E. Lee Day (Alabama, Arkansas, Florida, Georgia and Mississippi) Lee–Jackson–King Day (Virginia, United States, defunct) Husband's Day (Iceland) Kokborok Day (Tripura, India) National Popcorn Day (United States) Theophany / Epiphany (Eastern and Oriental Orthodoxy), and its related observances: Timkat, (on 20th during Leap Year) (Ethiopian Orthodox) Vodici or Baptism of Jesus (North Macedonia) January 20 Armed Forces Day (Mali) Cheese Day (United States) Heroes' Day (Cape Verde) Inauguration Day, held every four years in odd-numbered years, except when January 20 falls on a Sunday (United States) Martyrs' Day (Azerbaijan) January 21 Babinden (Bulgaria, Serbia) Birthday of Princess Ingrid Alexandra (Norway) Errol Barrow Day (Barbados) Flag Day (Quebec) Grandmother's Day (Poland) Lady of Altagracia Day (Dominican Republic) Lincoln Alexander Day (Canada) National Hug Day (United States) January 22 Answer Your Cat's Question Day Day of Unity of Ukraine (Ukraine) Grandfather's Day (Poland) National Hot Sauce Day (United States) January 23 Bounty Day (Pitcairn Island) Espousals of the Blessed Virgin Mary (Roman Catholic Church) National Pie Day (United States) Netaji Subhas Chandra Bose's Jayanti (Orissa, Tripura, and West Bengal, India) Snow Plow Mailbox Hockey Day World Freedom Day (Taiwan and South Korea) January 24 Feast of Our Lady of Peace (Roman Catholic Church), and its related observances: Feria de Alasitas (La Paz) Moebius Syndrome Awareness Day (international observance) National Peanut Butter Day (United States) Unification Day (Romania) January 25 2011 Revolution Day (Egypt) A Room of One's Own Day Burns night (Scotland, Scottish community) Dydd Santes Dwynwen (Wales) Feast of the Conversion of Saint Paul (Eastern Orthodox, Oriental Orthodox, Roman Catholic, Anglican and Lutheran churches, which concludes the Week of Prayer for Christian Unity) National Police Day (Egypt) National Voters' Day (India) Tatiana Day (Russia, Eastern Orthodox) January 26 Australia Day (Australia) Duarte Day (Dominican Republic) Engineer's Day (Panama) International Customs Day Liberation Day (Uganda) Republic Day (India) January 27 Day of the lifting of the siege of Leningrad (Russia) Liberation of the remaining inmates of Auschwitz-related observances: Holocaust Memorial Day (UK) Holocaust Remembrance Day (Sweden) International Holocaust Remembrance Day Memorial Day (Italy) Memorial Day for the Victims of the Holocaust and Prevention of Crimes against Humanity (Czech Republic) Memorial Day for the Victims of National Socialism (Germany) National Holocaust Memorial Day (Greece) Family Literacy Day (Canada) Feast of Saint Slava (Serbia) National Chocolate Cake Day (United States) Saint Devota's Day (Monaco) January 28 Army Day (Armenia) Data Privacy Day (international observance) January 29 Kansas Day (Kansas, United States) January 30 Day of Azerbaijani customs (Azerbaijan) Day of Saudade (Brazil) Fred Korematsu Day (California, United States) Martyrdom of Mahatma Gandhi-related observances: Martyrs' Day (India) School Day of Non-violence and Peace (Spain) Start of the Season for Nonviolence January 30 – April 4 National Inane Answering Message Day (United States) Teacher's Day (Greece) January 31 Amartithi (Meherabad, India, followers of Meher Baba) Independence Day (Nauru) Me-Dam-Me-Phi (Ahom people) Street Children's Day (Austria) References 01
15813
https://en.wikipedia.org/wiki/January%2030
January 30
Events Pre-1600 1018 – Poland and the Holy Roman Empire conclude the Peace of Bautzen. 1287 – King Wareru founds the Hanthawaddy Kingdom, and proclaims independence from the Pagan Kingdom. 1601–1900 1607 – An estimated 200 square miles (51,800 ha) along the coasts of the Bristol Channel and Severn Estuary in England are destroyed by massive flooding, resulting in an estimated 2,000 deaths. 1648 – Eighty Years' War: The Treaty of Münster and Osnabrück is signed, ending the conflict between the Netherlands and Spain. 1649 – Charles I of England is executed in Whitehall, London. 1661 – Oliver Cromwell, Lord Protector of the Commonwealth of England, is ritually executed more than two years after his death, on the 12th anniversary of the execution of the monarch he himself deposed. 1703 – The Forty-seven rōnin, under the command of Ōishi Kuranosuke, avenge the death of their master, by killing Kira Yoshinaka. 1789 – Tây Sơn forces emerge victorious against Qing armies and liberate the capital Thăng Long. 1806 – The original Lower Trenton Bridge (also called the Trenton Makes the World Takes Bridge), which spans the Delaware River between Morrisville, Pennsylvania and Trenton, New Jersey, is opened. 1820 – Edward Bransfield sights the Trinity Peninsula and claims the discovery of Antarctica. 1826 – The Menai Suspension Bridge, considered the world's first modern suspension bridge, connecting the Isle of Anglesey to the north West coast of Wales, is opened. 1835 – In the first assassination attempt against a President of the United States, Richard Lawrence attempts to shoot president Andrew Jackson, but fails and is subdued by a crowd, including several congressmen as well as Jackson himself. 1847 – Yerba Buena, California is renamed San Francisco, California. 1858 – The first Hallé concert is given in Manchester, England, marking the official founding of The Hallé orchestra as a full-time, professional orchestra. 1862 – The first American ironclad warship, the is launched. 1889 – Archduke Crown Prince Rudolf of Austria, heir to the Austro-Hungarian crown, is found dead with his mistress Baroness Mary Vetsera in the Mayerling. 1901–present 1902 – The first Anglo-Japanese Alliance is signed in London. 1908 – Indian pacifist and leader Mohandas Karamchand Gandhi is released from prison by Jan C. Smuts after being tried and sentenced to two months in jail earlier in the month. 1911 – The destroyer makes the first airplane rescue at sea saving the life of Douglas McCurdy from Havana, Cuba. 1920 – Japanese carmaker Mazda is founded, initially as a cork-producing company. 1925 – The Government of Turkey expels Patriarch Constantine VI from Istanbul. 1930 – The Politburo of the Communist Party of the Soviet Union orders the confiscation of lands belonging to the Kulaks in a campaign of Dekulakization, resulting in the executions and forced deportations of millions. 1933 – Adolf Hitler's rise to power: Hitler takes office as the Chancellor of Germany. 1939 – During a speech in the Reichstag, Adolf Hitler makes a prediction about the end of the Jewish race in Europe if another world war were to occur. 1942 – World War II: Japanese forces invade the island of Ambon in the Dutch East Indies. Some 300 captured Allied troops are killed after the surrender. One-quarter of the remaining POWs remain alive at the end of the war. 1944 – World War II: The Battle of Cisterna, part of Operation Shingle, begins in central Italy. 1945 – World War II: The Wilhelm Gustloff, overfilled with German refugees, sinks in the Baltic Sea after being torpedoed by a Soviet submarine, killing approximately 9,500 people. 1945 – World War II: Raid at Cabanatuan: One hundred and twenty-six American Rangers and Filipino resistance fighters liberate over 500 Allied prisoners from the Japanese-controlled Cabanatuan POW camp. 1948 – British South American Airways' Tudor IV Star Tiger disappears over the Bermuda Triangle. 1948 – Following the assassination of Mahatma Gandhi in his home compound, India's prime minister, Jawaharlal Nehru, broadcasts to the nation, saying "The light has gone out of our lives". The date of the assassination becomes observed as "Martyrs' Day" in India. 1956 – In the United States, Civil Rights Movement leader Martin Luther King Jr.'s home is bombed in retaliation for the Montgomery bus boycott. 1959 – The forces of the Sultanate of Muscat occupy the last strongholds of the Imamate of Oman, Saiq and Shuraijah, marking the end of Jebel Akhdar War in Oman. 1959 – , specifically designed to operate in icebound seas, strikes an iceberg on her maiden voyage and sinks, killing all 95 aboard. 1960 – The African National Party is founded in Chad, through the merger of traditionalist parties. 1964 – In a bloodless coup, General Nguyễn Khánh overthrows General Dương Văn Minh's military junta in South Vietnam. 1968 – Vietnam War: Tet Offensive launch by forces of the Viet Cong and North Vietnamese Army against South Vietnam, the United States, and their allies. 1969 – The Beatles' last public performance, on the roof of Apple Records in London. The impromptu concert is broken up by the police. 1972 – The Troubles: Bloody Sunday: British paratroopers open fire on anti-internment marchers in Derry, Northern Ireland, killing 13 people; another person later dies of injuries sustained. 1972 – Pakistan leaves the Commonwealth of Nations in protest of its recognition of breakaway Bangladesh. 1974 – Pan Am Flight 806 crashes near Pago Pago International Airport in American Samoa, killing 97. 1975 – The Monitor National Marine Sanctuary is established as the first United States National Marine Sanctuary. 1979 – A Varig Boeing 707-323C freighter, flown by the same commander as Flight 820, disappears over the Pacific Ocean 30 minutes after taking off from Tokyo. 1982 – Richard Skrenta writes the first PC virus code, which is 400 lines long and disguised as an Apple boot program called "Elk Cloner". 1989 – The American embassy in Kabul, Afghanistan is closed. 1995 – Hydroxycarbamide becomes the first approved preventive treatment for sickle cell disease. 2000 – Kenya Airways Flight 431 crashes into the Atlantic Ocean off the coast of Ivory Coast, killing 169. 2013 – Naro-1 becomes the first carrier rocket launched by South Korea. 2020 – The World Health Organization declares the COVID-19 pandemic to be a Public Health Emergency of International Concern. Births Pre-1600 58 BC – Livia, Roman wife of Augustus (d. 29) 1410 – William Calthorpe, English knight (d. 1494) 1520 – William More, English courtier (d. 1600) 1563 – Franciscus Gomarus, Dutch theologian and academic (d. 1641) 1573 – Georg Friedrich, Margrave of Baden-Durlach (d. 1638) 1580 – Gundakar, Prince of Liechtenstein, court official in Vienna (d. 1658) 1590 – Lady Anne Clifford, 14th Baroness de Clifford (d. 1676) 1601–1900 1628 – George Villiers, 2nd Duke of Buckingham, English statesman (d. 1687) 1661 – Charles Rollin, French historian and educator (d. 1741) 1697 – Johann Joachim Quantz, German flute player and composer (d. 1773) 1703 – François Bigot, French politician (d. 1778) 1720 – Charles De Geer, Swedish entomologist and archaeologist (d. 1778) 1754 – John Lansing, Jr., American lawyer and politician (d. 1829) 1775 – Walter Savage Landor, English poet and author (d. 1864) 1781 – Adelbert von Chamisso, German botanist and poet (d. 1838) 1816 – Nathaniel P. Banks, American general and politician, 24th Governor of Massachusetts (d. 1894) 1822 – Franz Ritter von Hauer, Austrian geologist and curator (d. 1899) 1841 – Félix Faure, French politician, 7th President of France (d. 1899) 1844 – Richard Theodore Greener, American lawyer, academic, and diplomat (d. 1922) 1846 – Angela of the Cross, Spanish nun and saint (d. 1932) 1861 – Charles Martin Loeffler, German-American violinist and composer (d. 1935) 1862 – Walter Damrosch, German-American conductor and composer (d. 1950) 1866 – Gelett Burgess, American author, poet, and critic (d. 1951) 1878 – A. H. Tammsaare, Estonian author (d. 1940) 1882 – Franklin D. Roosevelt, American lawyer and statesman, 32nd President of the United States (d. 1945) 1889 – Jaishankar Prasad, Indian poet and playwright (d. 1937) 1899 – Max Theiler, South African-American virologist and academic, Nobel Prize laureate (d. 1972) 1900 – Martita Hunt, Argentine-born British actress (d. 1969) 1901–present 1901 – Rudolf Caracciola, German racing driver (d. 1959) 1902 – Nikolaus Pevsner, German-English historian and scholar (d. 1983) 1910 – Chidambaram Subramaniam, Indian lawyer and politician, Indian Minister of Defence (d. 2000) 1911 – Roy Eldridge, American jazz trumpet player (d. 1989) 1912 – Werner Hartmann, German physicist and academic (d. 1988) 1912 – Francis Schaeffer, American pastor and theologian (d. 1984) 1912 – Barbara W. Tuchman, American historian and author (d. 1989) 1913 – Amrita Sher-Gil, Hungarian-Indian painter (d. 1941) 1914 – Luc-Marie Bayle, French commander and painter (d. 2000) 1914 – John Ireland, Canadian-American actor and director (d. 1992) 1914 – David Wayne, American actor (d. 1995) 1915 – Joachim Peiper, German SS officer (d. 1976) 1915 – John Profumo, English soldier and politician, Secretary of State for War (d. 2006) 1917 – Paul Frère, Belgian racing driver and journalist (d. 2008) 1918 – David Opatoshu, American actor and screenwriter (d. 1996) 1919 – Fred Korematsu, American activist (d. 2005) 1920 – Michael Anderson, English director and producer (d. 2018) 1920 – Patrick Heron, British painter (d. 1999) 1920 – Delbert Mann, American director and producer (d. 2007) 1922 – Dick Martin, American comedian, actor, and director (d. 2008) 1923 – Marianne Ferber, Czech-American economist and author (d. 2013) 1924 – S. N. Goenka, Burmese-Indian author and educator (d. 2013) 1924 – Lloyd Alexander, American soldier and author (d. 2007) 1925 – Douglas Engelbart, American computer scientist, invented the computer mouse (d. 2013) 1927 – Olof Palme, Swedish statesman, 26th Prime Minister of Sweden (d. 1986) 1928 – Harold Prince, American director and producer (d. 2019) 1929 – Lois Hole, Canadian businesswoman and politician, 15th Lieutenant Governor of Alberta (d. 2005) 1929 – Hugh Tayfield, South African cricketer (d. 1994) 1929 – Lucille Teasdale-Corti, Canadian-Italian physician and humanitarian (d. 1996) 1930 – Gene Hackman, American actor and author 1930 – Magnus Malan, South African general and politician, South African Minister of Defence (d. 2011) 1931 – John Crosbie, Canadian lawyer and politician, 34th Canadian Minister of Justice (d. 2020) 1931 – Shirley Hazzard, Australian-American novelist, short story writer, and essayist (d. 2016) 1932 – Knock Yokoyama, Japanese comedian and politician (d. 2007) 1934 – Tammy Grimes, American actress and singer (d. 2016) 1935 – Richard Brautigan, American novelist, poet, and short story writer (d. 1984) 1935 – Tubby Hayes, English saxophonist and composer (d. 1973) 1936 – Horst Jankowski, German pianist and composer (d. 1998) 1937 – Vanessa Redgrave, English actress 1937 – Boris Spassky, Russian chess player 1938 – Islam Karimov, Uzbek politician, 1st President of Uzbekistan (d. 2016) 1941 – Gregory Benford, American astrophysicist and author 1941 – Dick Cheney, American businessman and politician, 46th Vice President of the United States 1941 – Tineke Lagerberg, Dutch swimmer 1942 – Marty Balin, American singer-songwriter and guitarist (d. 2018) 1943 – Davey Johnson, American baseball player and manager 1944 – Lynn Harrell, American cellist and academic (d. 2020) 1944 – Colin Rimer, English lawyer and judge 1945 – Meir Dagan, Israeli military officer and intelligence official, Director of Mossad (2002–11) (d. 2016) 1945 – Michael Dorris, American author and scholar (d. 1997) 1946 – John Bird, Baron Bird, English publisher, founded The Big Issue 1947 – Les Barker, English poet and author 1947 – Steve Marriott, English singer-songwriter and guitarist (d. 1991) 1949 – Peter Agre, American physician and biologist, Nobel Prize laureate 1950 – Jack Newton, Australian golfer 1951 – Phil Collins, English drummer, singer-songwriter, producer, and actor 1951 – Charles S. Dutton, American actor and director 1951 – Bobby Stokes, English footballer (d. 1995) 1952 – Doug Falconer, Canadian football player and producer (d. 2021) 1953 – Fred Hembeck, American author and illustrator 1955 – John Baldacci, American politician, 73rd Governor of Maine 1955 – Curtis Strange, American golfer 1957 – Chris Jansing, American television reporter 1957 – Payne Stewart, American golfer (d. 1999) 1959 – Steve Folkes, Australian rugby league player and coach (d. 2018) 1959 – Jody Watley, American entertainer 1962 – Abdullah II of Jordan 1962 – Mary Kay Letourneau, American child rapist (d. 2020) 1964 – Otis Smith, American basketball player, coach, and manager 1965 – Kevin Moore, Australian rugby league player and coach 1966 – Danielle Goyette, Canadian ice hockey player and coach 1968 – Felipe VI of Spain 1969 – Justin Skinner, English footballer and manager 1970 – Kimiya Yui, Japanese astronaut 1973 – Jalen Rose, American basketball player and sportscaster 1974 – Christian Bale, British actor 1974 – Olivia Colman, English actress 1975 – Juninho Pernambucano, Brazilian footballer 1976 – Andy Milonakis, American entertainer 1978 – Carmen Küng, Swiss curler 1978 – John Patterson, American baseball player 1979 – Trevor Gillies, Canadian ice hockey player 1980 – João Soares de Almeida Neto, Brazilian footballer 1980 – Georgios Vakouftsis, Greek footballer 1980 – Wilmer Valderrama, American actor and producer 1981 – Dimitar Berbatov, Bulgarian footballer 1981 – Peter Crouch, English footballer 1981 – Mathias Lauda, Austrian racing driver 1982 – Jorge Cantú, Mexican baseball player 1984 – Kotoshōgiku Kazuhiro, Japanese sumo wrestler 1984 – Kid Cudi, American entertainer 1985 – Gisela Dulko, Argentinian tennis player 1987 – Becky Lynch, Irish wrestler 1987 – Renato Santos, Brazilian footballer 1987 – Arda Turan, Turkish footballer 1989 – Yoon Bo-ra, South Korean singer 1990 – Mitchell Starc, Australian cricketer 1990 – Phillip Supernaw, American football player 1991 – Stefan Elliott, Canadian ice hockey player 1993 – Katy Marchant, English track cyclist 1993 – Thitipoom Techaapaikhun, Thai actor 1995 – Jack Laugher, English diver 1996 – Dafne Navarro, Mexican trampoline gymnast Deaths Pre-1600 680 – Balthild, Frankish queen (b. 626) 970 – Peter I of Bulgaria 1030 – William V, Duke of Aquitaine (b. 969) 1181 – Emperor Takakura of Japan (b. 1161) 1240 – Pelagio Galvani, Leonese lawyer and cardinal (b. 1165) 1314 – Nicholas III of Saint Omer 1344 – William Montacute, 1st Earl of Salisbury (b. 1301) 1384 – Louis II, Count of Flanders (b. 1330) 1574 – Damião de Góis, Portuguese historian and philosopher (b. 1502) 1601–1900 1606 – Everard Digby, English criminal (b. 1578) 1606 – John Grant, English conspirator (b. 1570) 1606 – Robert Wintour, English conspirator (b. 1565) 1649 – Charles I of England, Scotland, and Ireland (b. 1600) 1664 – Cornelis de Graeff, Dutch mayor (b. 1599) 1730 – Peter II of Russia (b. 1715) 1770 – Giovanni Pietro Francesco Agius de Soldanis, Maltese linguist, historian and cleric (b. 1712) 1836 – Betsy Ross, American seamstress, said to have designed the American Flag (b. 1752) 1838 – Osceola, American tribal leader (b. 1804) 1858 – Coenraad Jacob Temminck, Dutch zoologist and ornithologist (b. 1778) 1867 – Emperor Kōmei of Japan (b. 1831) 1869 – William Carleton, Irish author (b. 1794) 1881 – Arthur O'Shaughnessy, English poet and herpetologist (b. 1844) 1889 – Rudolf, Crown Prince of Austria, heir apparent to the throne of Austria-Hungary (b. 1858) 1901–present 1926 – Barbara La Marr, American actress (b. 1896) 1928 – Johannes Fibiger, Danish physician and academic, Nobel Prize laureate (b. 1867) 1934 – Frank Nelson Doubleday, American publisher, founded the Doubleday Publishing Company (b. 1862) 1947 – Frederick Blackman, English botanist and physiologist (b. 1866) 1948 – Arthur Coningham, Australian air marshal (b. 1895) 1948 – Mahatma Gandhi, leader of the Indian independence movement against British rule (b. 1869) 1948 – Orville Wright, American pilot and engineer, co-founded the Wright Company (b. 1871) 1951 – Ferdinand Porsche, Austrian-German engineer and businessman, founded Porsche (b. 1875) 1958 – Jean Crotti, Swiss painter (b. 1878) 1958 – Ernst Heinkel, German engineer and businessman; founded the Heinkel Aircraft Company (b. 1888) 1962 – Manuel de Abreu, Brazilian physician and engineer (b. 1894) 1963 – Francis Poulenc, French pianist and composer (b. 1899) 1966 – Jaan Hargel, Estonian flute player, conductor, and educator (b. 1912) 1968 – Makhanlal Chaturvedi, Indian poet, playwright, and journalist (b. 1889) 1969 – Dominique Pire, Belgian friar, Nobel Prize laureate (b. 1910) 1973 – Elizabeth Baker, American economist and academic (b. 1885) 1974 – Olav Roots, Estonian pianist and composer (b. 1910) 1977 – Paul Marais de Beauchamp, French zoologist (b. 1883) 1980 – Professor Longhair, American singer-songwriter and pianist (b. 1918) 1982 – Lightnin' Hopkins, American singer-songwriter and guitarist (b. 1912) 1991 – John Bardeen, American physicist and engineer, Nobel Prize laureate (b. 1908) 1991 – Clifton C. Edom, American photographer and educator (b. 1907) 1993 – Alexandra of Yugoslavia, the last Queen of Yugoslavia (b. 1921) 1994 – Pierre Boulle, French soldier and author (b. 1912) 1999 – Huntz Hall, American actor (b. 1919) 1999 – Ed Herlihy, American journalist (b. 1909) 2001 – Jean-Pierre Aumont, French soldier and actor (b. 1911) 2001 – Johnnie Johnson, English air marshal and pilot (b. 1915) 2001 – Joseph Ransohoff, American surgeon and educator (b. 1915) 2004 – Egon Mayer, Swiss-American sociologist (b. 1944) 2005 – Martyn Bennett, Canadian-Scottish violinist (b. 1971) 2006 – Coretta Scott King, American author and activist (b. 1927) 2006 – Wendy Wasserstein, American playwright and academic (b. 1950) 2007 – Sidney Sheldon, American author and screenwriter (b. 1917) 2008 – Marcial Maciel, Mexican-American priest, founded the Legion of Christ and Regnum Christi (b. 1920) 2009 – H. Guy Hunt, American soldier, pastor, and politician, 49th Governor of Alabama (b. 1933) 2010 – Fadil Ferati, Kosovar accountant and politician (b. 1960) 2011 – John Barry, English composer and conductor (b. 1933) 2012 – Frank Aschenbrenner, American football player and soldier (b. 1925) 2012 – Doeschka Meijsing, Dutch author (b. 1947) 2013 – Gamal al-Banna, Egyptian author and scholar (b. 1920) 2013 – Patty Andrews, American singer (b. 1918) 2013 – George Witt, American baseball player and coach (b. 1931) 2014 – Stefan Bałuk, Polish general and photographer (b. 1914) 2014 – The Mighty Hannibal, American singer-songwriter and producer (b. 1939) 2014 – William Motzing, American composer and conductor (b. 1937) 2014 – Arthur Rankin, Jr., American director, producer, and screenwriter (b. 1924) 2015 – Carl Djerassi, Austrian-American chemist, author, and playwright (b. 1923) 2015 – Ülo Kaevats, Estonian academic, philosopher, and politician (b. 1947) 2015 – Geraldine McEwan, English actress (b. 1932) 2015 – Gerrit Voorting, Dutch cyclist (b. 1923) 2015 – Zhelyu Zhelev, Bulgarian philosopher and politician, 2nd President of Bulgaria (b. 1935) 2016 – Frank Finlay, English actor (b. 1926) 2016 – Francisco Flores Pérez, Salvadorian politician, President of El Salvador (b. 1959) 2016 – Georgia Davis Powers, American activist and politician (b. 1923) 2018 – Mark Salling, American actor and musician (b. 1982) 2019 – Dick Miller, American actor (b. 1928) 2021 – Sophie, Scottish musician (b. 1986) 2022 – Cheslie Kryst, American television presenter and model (b. 1991) 2023 – Bobby Beathard, American Pro Football Hall of Fame executive (b. 1937) 2023 – Bobby Hull, Canadian ice hockey player (b. 1939) Holidays and observances Christian Feast Day: Adelelmus of Burgos Aldegonde Anthony the Great (Coptic Church) Armentarius of Pavia Balthild Charles, King and Martyr (various provinces of the Anglican Communion) Hippolytus of Rome Hyacintha Mariscotti Martina Matthias of Jerusalem Mutien-Marie Wiaux Savina Three Holy Hierarchs (Eastern Orthodox), and its related observances: Teacher's Day (Greece) January 30 (Eastern Orthodox liturgics) Day of Azerbaijani customs (Azerbaijan) Day of Saudade (Brazil) Fred Korematsu Day (California, Florida, Hawaii, Virginia) Martyrdom of Mahatma Gandhi, and its related observances: Martyrs' Day (India) School Day of Non-violence and Peace (Spain) Start of the Season for Nonviolence (January 30 – April 4) References External links BBC: On This Day Historical Events on January 30 Days of the year January
15949
https://en.wikipedia.org/wiki/James%20Blish
James Blish
James Benjamin Blish () was an American science fiction and fantasy writer. He is best known for his Cities in Flight novels and his series of Star Trek novelizations written with his wife, J. A. Lawrence. His novel A Case of Conscience won the Hugo Award. He is credited with creating the term "gas giant" to refer to large planetary bodies. Blish was a member of the Futurians. His first published stories appeared in Super Science Stories and Amazing Stories. Blish wrote literary criticism of science fiction using the pen name William Atheling Jr. His other pen names included Donald Laverty, John MacDougal, and Arthur Lloyd Merlyn. Life Blish was born on May 23, 1921, at East Orange, New Jersey. While in high school, Blish self-published a fanzine, called The Planeteer, using a hectograph. The fanzine ran for six issues. Blish attended meetings of the Futurian Science Fiction Society in New York City during this period. Futurian members Damon Knight and Cyril M. Kornbluth became close friends. However, Blish's relationships with other members were often bitter. A personal target was fellow member Judith Merril, with whom he would debate politics. Merril would frequently dismiss Blish's self-description of being a "paper fascist". She wrote in Better to Have Loved (2002), "Of course [Blish] was not fascist, antisemitic, or any of those terrible things, but every time he used the phrase, I saw red." Blish studied microbiology at Rutgers University, graduating in 1942. He was drafted into Army service, and he served briefly as a medical laboratory technician. The United States Army discharged him for refusing orders to clean a grease trap in 1944. Following discharge, Blish entered Columbia University as a masters student of zoology. He did not complete the program, opting to write fiction full-time. In 1947, he married Virginia Kidd, a fellow Futurian. They divorced in 1963. Blish then married artist J. A. Lawrence in 1964, moving to England that same year. From 1962 to 1968, Blish worked for the Tobacco Institute as a writer and critic. Much of his work for the institute went uncredited. Blish died on July 30, 1975 from complications related to lung cancer. He was buried in Holywell Cemetery, Oxford. The Bodleian Library at Oxford is the custodian of Blish's papers. The library also has a complete catalog of Blish's published works. Career Throughout the 1940s, Blish published most of his stories in the few pulp magazines still in circulation. His first story was sold to fellow Futurian Frederik Pohl for Super Science Stories (1940), called "Emergency Refueling". Other stories were published intermittently, but with little circulation. Blish's "Chaos, Co-Ordinated", co-written with Robert A. W. Lowndes, was sold to Astounding Science Fiction, appearing in the October 1946 issue, earning Blish national circulation for the first time. Pantropy (1942–1956) Blish was what Andrew Litpack called a "practical writer". He would revisit, revise, and often expand on previously written stories. An example is "Sunken Universe" published in Super Science Stories in 1942. The story reappeared in Galaxy Science Fiction as "Surface Tension", in an altered form in 1952. The premise emphasized Blish's understanding of microbiology, and featured microscopic humans engineered to live on a hostile planet's shallow pools of water. The story proved to be among Blish's more popular and was anthologized in the first volume of Robert Silverberg The Science Fiction Hall of Fame. It was also anthologized in The Big Book of Science Fiction (2016), edited by Ann and Jeff VanderMeer. The world of microscopic humans continued in "The Thing in the Attic" in 1954, and "Watershed" the following year. The fourth entry, "A Time to Survive", was published by The Magazine of Fantasy & Science Fiction in 1957. The stories were collected, edited together, and published as the fix-up The Seedling Stars (1956), by Gnome Press. John Clute said all of Blish's "deeply felt work" explored "confronting the Faustian (or Frankensteinian) man". Cities in Flight (1950–1958) The Encyclopedia of Science Fiction asserts that not until the 1950s, and the Okie sequence of stories beginning their run, "did it become clear [Blish] would become a [science fiction] writer of unusual depth". The stories were loosely based on the Okie migration following the Dust Bowl of the 1930s, and were influenced by Oswald Spengler's two-part Der Untergang des Abendlandes (The Decline of the West). The stories detail the life of the Okies, humans who migrate throughout space looking for work in vast city-ships, powered by spindizzies, a type of anti-gravity engine. The premise and plot reflected Blish's feelings on the state of western civilization, and his personal politics. The first two stories, "Okie", and "Bindlestiff", were published in 1950, by Astounding. "Sargasso of Lost Cities" appeared in Two Complete Science-Adventure Books in April 1953. "Earthman, Come Home" followed a few months later, published by Astounding. In 1955, Blish collected the four stories together into an omnibus titled Earthman, Come Home, published by Putnam. More stories followed: In 1956, They Shall Have Stars, which edited together "Bridge" and "At Death’s End", and in 1958, Blish published The Triumph of Time. Four years later, he published a new Okies novel, A Life for the Stars. The Okies sequence was edited together and published as Cities In Flight (1970). Clute notes, "the brilliance of Cities in Flight does not lie in the assemblage of its parts, but in the momentum of the ideas embodied in it (albeit sometimes obscurely)." After Such Knowledge (1958–1971) Blish continued to rework older stories, and did so for one of his best known works, A Case of Conscience (1958). The novel originated as a novella, originally published in an issue of If, in 1953. The story follows a Jesuit priest, Ramon Ruiz-Sanchez, who visits the planet Lithia as a technical member of an expedition. While on the planet they discover a race of bipedal reptilians that have perfected morality in what Ruiz-Sanchez says is "the absence of God", and theological complications ensue. The book is one of the first major works in the genre to explore religion and its implications. It was the first of a series including Doctor Mirabilis (1964) and the two-part story Black Easter (1968), and The Day After Judgment (1971). The latter two were collected as The Devil's Day (1980). An omnibus of all four entries in the series was published by Legend in 1991, titled After Such Knowledge. A Case of Conscience won the 1959 Hugo Award for Best Novel, and was collected as part of Library of America’s omnibus American Science Fiction: Five Classic Novels 1956-1958. Star Trek (1967–1977) Bantam Books commissioned Blish to adapt episodes of Star Trek. The adapted short stories were generally based on draft scripts and contained different plot elements from the aired television episodes. The stories were collected into twelve volumes and published as a title series of the same name from 1967 to 1977. The adaptations were largely written by Blish; however, his declining health during this period proved problematic. His wife, J. A. Lawrence, wrote a number of installments. Her work remained uncredited until the final volume, Star Trek 12, published in 1977, two years after Blish's death. The first original novel for adults based on the television series, Spock Must Die! (1970), was also written by Blish, and he planned to release more. According to Lawrence, two episodes featuring popular character Harry Mudd, "I, Mudd" and "Mudd's Women", were held back by Blish for adaptation to be included in the follow-up to Spock Must Die!. However, Blish died before a novel could be completed. Lawrence did eventually adapt the two episodes, as Mudd's Angels (1978), which included an original novella The Business, as Usual, During Altercations by Lawrence. In her introduction to Star Trek 12, Lawrence states that Blish "did indeed write" adaptations of the two episodes. The introduction to Mudd's Angels acknowledges this, stating that Blish left the two stories in various stages of completion and they were finished by Lawrence; Blish does not receive author credit on the book. Blish credited his financial stability later in life to the Star Trek commission and the advance he received for Spock Must Die!. Literary criticism and legacy Blish was among the first literary critics of science fiction, and he judged works in the genre by the standards applied to "serious" literature. He took to task his fellow authors for deficiencies, such as bad grammar and a misunderstanding of scientific concepts, and the magazine editors, who accepted and published such material without editorial intervention. His critiques were published in "fanzines" in the 1950s under the pseudonym William Atheling Jr. The essays were collected in The Issue at Hand (1964) and More Issues at Hand (1970). Reviewing The Issue at Hand, Algis Budrys said that Atheling had, along with Damon Knight, "transformed the reviewer's trade in this field". He described the persona of Atheling as "acidulous, assertive, categorical, conscientious and occasionally idiosyncratic". Blish was a fan of the works of James Branch Cabell, and for a time edited Kalki, the journal of the Cabell Society. In his works of science fiction, Blish developed many ideas and terms which have influenced other writers and on occasion have been adopted more widely, such as faster than light communication via the dirac communicator, introduced in the short story "Beep" (1954). The dirac is comparable to Ursula K. Le Guin's ansible. Blish is also credited with coining the term gas giant, first used in the story "Solar Plexus", collected in the anthology Beyond Human Ken, edited by Judith Merril. The story was originally published in 1941, but did not contain the term. Blish reworked the story, changing the description of a large magnetic field to "a magnetic field of some strength nearby, one that didn't belong to the invisible gas giant revolving half a million miles away". Honors, awards and recognition The British Science Fiction Foundation inaugurated the James Blish Award for science fiction criticism in 1977, recognizing Brian W. Aldiss. The Science Fiction and Fantasy Hall of Fame inducted him in 2002. Awards and nominations 1959 Hugo Award for Best Novel, for A Case of Conscience. 1965 Nebula Award nomination for Best Novelette, for "The Shipwrecked Hotel", with Norman L. Knight. 1968 Nebula Award nomination for Best Novel, for Black Easter. 1970 Hugo Award nomination for Best Novella, for We All Die Naked. 1970 Nebula Award nomination for Best Novella, for A Style in Treason. Posthumous Awards and nominations 2001 [1951] Retro-Hugo Award nomination for Best Novelette, for "Okie". 2004 [1954] Retro-Hugo Award for Best Novella, for A Case of Conscience. 2004 [1954] Retro-Hugo Award for Best Novelette, for "Earthman, Come Home". Guest of Honor 1960 Guest of Honor, 18th World Science Fiction Convention. 1970 Guest of Honor, Scicon 70. Bibliography Blish's work was published by a variety of publishers in the United Kingdom and the United States, often with variations between editions, and with different titles. Blish also expanded and re-published his older work on numerous occasions. His works continued to be re-published after his death. Note: Very few of Blish's first editions were assigned ISBN numbers. Short fiction and novellas (1935–1986) Novels published in complete form, or serialized, in fiction magazines are included for completeness, and to avoid confusion. Novelette, Novella, Novel. The Planeteer (1935–1936) "Neptunian Refuge" (November 1935) "Mad Vision" (December 1935) "Pursuit into Nowhere" (January 1936) "Threat from Copernicus" (February 1936) "Trail of the Comet" (March 1936) "Bat-Shadow Shroud" (April 1936) Super Science Stories (1940) "Emergency Refueling" (March 1940) "Bequest of the Angel" (May 1940) "Sunken Universe" (May 1942),rewritten as "Surface Tension" (1952) Stirring Science Stories (1941) "Citadel of Thought" (February 1941) "Callistan Cabal" (April 1941) Science Fiction Quarterly (1941) "Weapon Out of Time" (April 1941) "When Anteros Came" (December 1941) Cosmic Stories (1941) "Phoenix Planet" (May 1941) "The Real Thrill" (July 1941) Future (1941–1953) "The Topaz Gate" (August 1941) "The Solar Comedy" (June 1942) "The Air Whale" (August 1942) "Struggle in the Womb" (May 1950) "The Secret People" (November 1950) "Elixir" (September 1951) "Testament of Andros" (January 1953) Astonishing Stories (1941) "Solar Plexus" (September 1941) Super Science and Fantastic Stories (1944) "The Bounding Crown" (December 1944) Science*Fiction (1946) "Knell", as by Arthur Lloyd Merlyn (January 1946) Astounding Science Fiction (1946–1957) "Chaos, Co-Ordinated" as by John MacDougal, with Robert A. W. Lowndes (October 1946) "Tiger Ride" with Damon Knight (October 1948) "Okie" (April 1950) "Bindlestiff" (December 1950) "Bridge" (February 1952) "Earthman, Come Home" (November 1953) "At Death's End" (May 1954) "One-Shot" (August 1955) "Tomb Tapper" (July 1956) Get Out of My Sky (January 1957),included in Get out of My Sky Panther ed. (1980) Startling Stories (1948) "Mistake Inside" (April 1948) Planet Stories (1948–1951) "Against the Stone Beasts" (August 1948) "Blackout in Cygni" (July 1951) Thrilling Wonder Stories (1948–1950) "No Winter, No Summer" as by Donald Laverty, with Damon Knight (October 1948) "The Weakness of RVOG" (February 1949),expanded as VOR (1958) "The Box" (April 1949) "The Homesteader" (June 1949) Let the Finder Beware (December 1949) "There Shall Be No Darkness" (April 1950),included in Get Out of My Sky Panther ed. (1980) Jungle Stories (1948) "Serpent's Fetish" (December 1948) Fantastic Story Quarterly (1950) "The Bore" (July 1950) Imagination (1951) "The Void Is My Coffin" (June 1951) Two Complete Science-Adventure Books (1951) The Warriors of Day (August 1951) Sargasso of Lost Cities (April 1953) Other Worlds Science Stores (1952) "Nightride and Sunrise" with Jerome Bixby (June 1952) Galaxy Science Fiction (1952–1970) "Surface Tension" (August 1952),collected in The Seedling Stars (1957) "Beep" (February 1954),expanded as The Quincunx of Time (1973) "The Writing of the Rat" (July 1956) "The Genius Heap" (August 1956) "On the Wall of the Lodge" with Virginia Kidd (June 1962) "The Shipwrecked Hotel" with Norman L. Knight, (August 1965),expanded as A Torrent of Faces (1967) "The Piper of Dis" with Norman L. Knight, (August 1966),expanded as A Torrent of Faces (1967) "Our Binary Brothers" (February 1969) "The City That Was the World" (July 1969) "A Style in Treason" (May 1970) The Day After Judgment (September 1970),collected in The Devil's Day (1990) "Darkside Crossing" (December 1970) "The Glitch" (June 1974) "The Art of the Sneeze" (November 1982) Dynamic Science Fiction (1953) "Turn of a Century" (March 1953) The Duplicated Man with Robert A. W. Lowndes (August 1953) Worlds of If (1953–1968) A Case of Conscience (September 1953),expanded as A Case of Conscience (1958) "The Thing in the Attic" (July 1954),collected in The Seedling Stars (1957) "Watershed" (May 1955),collected in The Seedling Stars (1957) "To Pay the Piper" (February 1956) Welcome to Mars (July 1966) Black Easter (August 1967),collected in The Devil's Day (1990) "Now That Man Is Gone" (November 1968) Star Science Fiction Stories (1953) "F.Y.I." (December 1953) The Magazine of Science Fiction and Fantasy (1953–1980) "First Strike" (June 1953) "The Book of Your Life" (March 1955) "With Malice to Come (3 vignettes)" (May 1955) "A Time to Survive" (February 1956),collected in The Seedling Stars Signet ed. (1959) "This Earth of Hours" (June 1959) "The Masks" (November 1959) "The Oath" (October 1960) "Who's in Charge Here?" (May 1962) "No Jokes on Mars" (October 1965) Midsummer Century (November 1982) Fantastic Universe (1955) "Translation" (March 1955) Infinity Science Fiction (1955–1957) "King of the Hill" (November 1955) "Sponge Dive" (June 1956) "Detour to the Stars" (December 1956) "Nor Iron Bars" (November 1957),expanded as Galactic Cluster (1959) Science Fiction Stories (1956) "A Work of Art" (July 1956) Science Fiction Adventures (1957) Two Worlds in Peril (February 1957) Amazing Stories (1960–1961) … And All the Stars a Stage (June 1960) "And Some Were Savages" (November 1960) "A Dusk of Idols" (March 1961) Impulse (1966) "A Hero's Life" (March 1966) Analog (1967–1968) "To Love Another" (April 1967),expanded as A Torrent of Faces (1967) "Skysign" with Norman L. Knight, (May 1968) Penthouse (1972) "A Light to Fight by" (June 1972) Fantasy Book (1986) "The White Empire" (September 1986) Anthologized short fiction (1952–2008) Beanstalk, Future Tense (1952), ed. Kendell Foster Crossen. Greenberger.expanded in Titan's Daughter (1961). "Common Time", Shadows of Tomorrow (August 1953), ed. Frederik Pohl. Permabooks #P236. "A Matter of Energy", The Best from Fantasy and Science Fiction, Fifth Series (January 1956), ed. Anthony Boucher. Doubleday. "Nor Iron Bars" (expanded), Galactic Cluster (October 1959), ed. James Blish. Signet #S1719. "The Abattoir Effect", So Close to Home (February 27, 1961), ed. James Blish. Ballantine Books #465K. "None So Blind", Anywhen (July 1970), ed. James Blish. Doubleday. "How Beautiful With Banners", Orbit 1 (1966), ed. Damon Knight. Whiting & Wheaton. "We All Die Naked", Three for Tomorrow (August 1969), ed. uncredited. Meredith Press. "More Light", Alchemy and Academe (November 1970), ed. Anne McCaffrey. Doubleday. "Statistician's Day", Science Against Man (December 1970), ed. Anthony Cheetham. Avon #V2374. "Getting Along", Again, Dangerous Visions (March 17, 1972), ed. Harlan Ellison. Doubleday. "A True Bill: A Chancel Drama in One Act", Ten Tomorrows (September 1973), ed. Roger Elwood. Fawcett Gold Medal #M2820. "The Price of a Drink", The Berserkers (January 1974), ed. Roger Elwood. Trident . "Making Waves", Works of Art (January 30, 2008). NESFA Press . Novels (1952–1990) Jack of Eagles (1952). Greenberg.,also published as ESPer (1952). Avon. The Frozen Year (March 19, 1957). Ballantine Books #197,also published as Fallen Star (1957). Faber & Faber. VOR (April 1958). Avon #T-238. The Duplicated Man (1959). Avalon Books. A Torrent of Faces (1967), with Norman L. Knight. Doubleday. The Warriors of Day (1967). Lancer Books #73-580. The Star Dwellers (1961). G. P. Putnam's Sons. Titan's Daughter (March 1961). Berkley Medallion #G507. The Night Shapes (October 1962). Ballantine Books #F647. Mission to the Heart Stars (November 11, 1965). Faber & Faber. Welcome to Mars (July 1966). G. P. Putnam's Sons. The Vanished Jet (1968). Weybright and Talley. … And All the Stars a Stage (July 1971). Doubleday. Midsummer Century (May 1972). Doubleday,included in Midsummer Century Daw ed. (1974). The Quincunx of Time (October 1973). Dell #07244. Cities in Flight series (1955–1962) Earthman, Come Home (1955). G. P. Putnam's Sons. They Shall Have Stars (1956). Faber & Faber,also published as Year 2018! (1957). Avon Books. The Triumph of Time (October 1958). Avon #T-279,also published as A Clash of Cymbals (1959). Faber & Faber. A Life for the Stars (1962). G. P. Putnam's Sons. After Such Knowledge series (1958–1990) A Case of Conscience (April 1958). Ballantine Books #256. Doctor Mirabilis (1964). Faber & Faber #55198. The Devil's Day (February 1990). Baen . Collections (1957–2009) The Seedling Stars (1957). Gnome Press. The Seedling Stars (February 1959). Signet #S1622. Best Science Fiction Stories of James Blish (1965). Faber & Faber,also published as The Testament of Andros (August 1977). Arrow Books . Midsummer Century (February 1974). Daw #UQ1094. The Best of James Blish (August 1979). Ballantine/Del Rey Get Out of My Sky (April 1980). Panther . A Work of Art and Other Stories (July 1993). Severn House . With All Love: Selected Poems (March 1995). Anamnesis Press . A Dusk of Idols and Other Stories (May 1996). Severn House . In This World, or Another (July 2, 2003). Five Star . Works of Art (January 30, 2008). NESFA Press . Flights of Eagles (October 20, 2009). NESFA Press . Anthologies (1959–1970) Galactic Cluster (October 1959). Signet #S1719. So Close to Home (February 27, 1961). Ballantine Books #465K. New Dreams This Morning (October 1966). Ballantine Books #U233. Anywhen (1970). Doubleday. Nebula Award Stories 5 (1970). Gollancz. Nonfiction (1964–1987) The Issue at Hand (1964), as by William Atheling Jr. Advent Publishers. More Issues at Hand (December 1970), as by William Atheling Jr. Advent Publishers . The Tale That Wags the God (July 1987). Advent Publishers . Star Trek (1967–1977) Star Trek (January 1967). Bantam Books #F3459. Star Trek 2 (February 1968). Bantam Books #F3439. Star Trek 3 (April 1969). Bantam Books #F4371. Spock Must Die! (February 1970). Bantam Books #H5515. Star Trek 4 (July 1971). Bantam Books #S7009. Star Trek 5 (February 1972). Bantam Books #S7300. Star Trek 6 (April 1972). Bantam Books #S7364. Star Trek 7 (July 1972). Bantam Books #S7480. Star Trek 8 (November 1972). Bantam Books #SP7550. Star Trek 9 (August 1973). Bantam Books #SP7808. Star Trek 10 (February 1974). Bantam Books #SP8401. Star Trek 11 (April 1975). Bantam Books #Q8717,also published as The Day of the Dove (October 1985). Spectra . Star Trek 12 (November 1977), with J. A. Lawrence. Bantam Books . Omnibuses (1970–2013) Cities in Flight (February 1970). Avon #W187. After Such Knowledge (July 1991). Legend ). The Seedling Stars / Galactic Cluster (April 1983). Signet . Black Easter / The Day After Judgement / The Seedling Stars (September 26, 2013) . References Further reading "Special James Blish Issue" (April 1972) The Magazine of Fantasy & Science Fiction. . External links Audio of Blish giving the Guest of Honor speech at Pittcon 1960 Audio of Blish giving the Guest of Honour speech at Eastercon 1970 1921 births 1975 deaths 20th-century American essayists 20th-century American male writers 20th-century American novelists 20th-century American short story writers American fantasy writers American literary critics American male essayists American male novelists American male short story writers American science fiction writers American speculative fiction critics American speculative fiction writers American anthologists Burials at Holywell Cemetery Columbia Graduate School of Arts and Sciences alumni Deaths from lung cancer in England Faster-than-light communication Futurians Hugo Award-winning writers Literacy and society theorists Literary theorists Novelists from New Jersey Science fiction critics Science Fiction Hall of Fame inductees Star Trek fiction writers Rutgers University alumni Theorists on Western civilization United States Army personnel of World War II Writers about religion and science Writers from East Orange, New Jersey 20th-century pseudonymous writers
16111
https://en.wikipedia.org/wiki/John%20Sayles
John Sayles
John Thomas Sayles (born September 28, 1950) is an American independent film director, screenwriter, editor, actor, and novelist. He has twice been nominated for the Academy Award for Best Original Screenplay, for Passion Fish (1992) and Lone Star (1996). His film Men with Guns (1997) was nominated for the Golden Globe Award for Best Foreign Language Film. His directorial debut, Return of the Secaucus 7 (1980), was added to the National Film Registry in 1997. Early life Sayles was born on September 28, 1950, in Schenectady, New York, the son of Mary (née Rausch), a teacher, and Donald John Sayles, a school administrator. Both of Sayles's parents were Catholic and of half-Irish descent. Sayles has referred to himself as a "Catholic atheist". He attended Williams College with frequent collaborators Gordon Clapp and David Strathairn, as well as his longtime partner, Maggie Renzi. Sayles earned a B.A. in psychology in 1972. Career After college, Sayles moved to Boston where he worked a variety of blue-collar jobs while writing short stories for The Atlantic. These writings culminated in his first novel, The Pride of the Bimbos, published in 1975. Like Martin Scorsese and Francis Ford Coppola, Sayles began his film career working with Roger Corman. In 1979, Sayles used $30,000 he earned writing scripts for Corman to fund his first film, Return of the Secaucus 7. To make the film on a limited budget, he set the film in a large house so that he did not have to travel to or get permits for different locations, set the story over a three-day weekend to limit costume changes, and wrote about people his age so he could cast his friends in it. The film received near-unanimous critical acclaim at the time and has held its reputation. In November 1997, the National Film Preservation Board announced that Return of the Secaucus 7 would be one of the 25 films selected that year for preservation in the National Film Registry at the Library of Congress. In 1983, after the films Baby It's You (starring Rosanna Arquette) and Lianna (a story in which a married woman becomes discontented with her marriage and falls in love with another woman), Sayles received a MacArthur Fellowship. He put the money into the science fiction feature The Brother from Another Planet, a film about a three-toed humanoid who escapes bondage on another world and crash-lands in New York harbour; because he is Africanoid in appearance, he finds himself at home among the people of Harlem, being pursued by European-looking alien enslavers men in black. In 1989, Sayles created and wrote the pilot episode for the short-lived television show Shannon's Deal about a down-and-out Philadelphia lawyer played by Jamey Sheridan. Sayles received a 1990 Edgar Award for his teleplay for the pilot. The show ran for 16 episodes before being cancelled in 1991. Sayles has funded most of his films by writing genre scripts, such as Piranha, Alligator, The Howling, and The Challenge Having collaborated with Joe Dante on Piranha and The Howling, Sayles acted in Dante's movie, Matinee. Sayles gets the rest of his funding by working as a script doctor; he did rewrites for Apollo 13 and Mimic. A genre script, called Night Skies, inspired what would eventually become the film E.T. the Extra-Terrestrial. That film's director, Steven Spielberg, later commissioned Sayles to write a script (unused) for the fourth Jurassic Park film. He has written and directed his own films, including Lone Star, Passion Fish, Eight Men Out, The Secret of Roan Inish, and Matewan. He serves on the advisory board for the Austin Film Society. Maggie Renzi has been John Sayles's long-time companion (and collaborator), but they have not married. Renzi has produced most of his films since Lianna. They met as students at Williams College. Sayles works with a regular repertory of actors, most notably Chris Cooper, David Strathairn, and Gordon Clapp, each of whom has appeared in at least four of his films. In early 2003, Sayles signed the Not In Our Name "Statement of Conscience" (along with Noam Chomsky, Steve Earle, Brian Eno, Jesse Jackson, Viggo Mortensen, Bonnie Raitt, Oliver Stone, Marisa Tomei, Susan Sarandon and others) which opposed the invasion of Iraq. In February 2009, Sayles was reported to be writing an HBO series based on the early life of Anthony Kiedis of the Red Hot Chili Peppers. The drama, tentatively titled Scar Tissue, centers on Kiedis's early years living in West Hollywood with his father. At that time, Kiedis's father, known as Spider, sold drugs (according to legend, his clients included The Who and Led Zeppelin) and mingled with rock stars on the Sunset Strip, all while aspiring to get into show business. In February 2010, Sayles began shooting his 17th feature film, the historical war drama Amigo, in the Philippines. The film is a fictional account of events during the Philippine–American War, with a cast that includes Joel Torre, Chris Cooper, and Garret Dillahunt. His novel A Moment in the Sun, set during the same period as Amigo, in the Philippines, Cuba, and the U.S., was released in 2011 by McSweeney's. It includes an account of the Wilmington Insurrection of 1898 in North Carolina, the only coup d'état in United States history in which a duly elected government was overthrown. Legacy and honors 1983 MacArthur Fellowship 1990 Edgar Award, for teleplay for pilot of Shannon's Deal In June 2014, Sayles donated his non-film archive to the University of Michigan. It will be accessible at the Harlan Hatcher Graduate Library. Sayles's film archive is held by the UCLA Film and Television Archive. Filmography Writer/director Return of the Secaucus 7 (1980) Lianna (1983) Baby It's You (1983) The Brother from Another Planet (1984) Matewan (1987) Eight Men Out (1988) (also actor portraying Ring Lardner) City of Hope (1991) Passion Fish (1992) The Secret of Roan Inish (1994) Lone Star (1996) Men with Guns (1997) Limbo (1999) Sunshine State (2002) Casa de los Babys (2003) Silver City (2004) Honeydripper (2007) Amigo (2010) Go for Sisters (2013) Writer (film) Piranha (1978) The Lady in Red (1979) Battle Beyond the Stars (1980) Alligator (1980) The Howling (1981) The Challenge (with Richard Maxwell) (1982) E.T. the Extra Terrestrial (early draft, then titled Night Skies) Enormous Changes at the Last Minute (with Susan Rice) (1983) The Clan of the Cave Bear (1986) Wild Thing (1987) Breaking In (1989) Men of War (as A Safe Place, later repolished by Ethan Reiff and Cyrus Voris) (1994) Apollo 13 (1995) (uncredited rewrite) The Spiderwick Chronicles (Co-writer with David Berenbaum and Karey Kirkpatrick) (2008) The Devil's Highway (2018) Writer (TV) A Perfect Match (with Mel Damski) (1980) Unnatural Causes (1986) Shannon's Deal (1989) (Creator) The Alienist (2018) Actor (film) Return of the Secaucus 7 (as Howie) (1980) Lianna (as Jerry) (1983) The Brother from Another Planet (as Man in Black #2) (1984) Something Wild (as Motorcycle Cop) (1986) Matewan (as Hardshell Preacher) (1987) Eight Men Out (as Ring Lardner) (1988) City of Hope (as Carl) (1991) Matinee (as Bob) (1993) Gridlock'd (1996) In the Electric Mist (as Michael Goldman) (2009) The Normals (as Dr. Marx) (2012) Bibliography Novels Pride of the Bimbos (1975) (novel) Union Dues (1977) (novel) Los Gusanos (1991) (novel) A Moment in the Sun (2011) (novel) Yellow Earth (2020) (novel) Collections and non-fiction The Anarchists' Convention (1979) (short story collection) Thinking in Pictures: The Making of the Movie "Matewan" (1987) (non-fiction) Dillinger in Hollywood (2004) (short story collection) Music videos Bruce Springsteen – "Born in the U.S.A." Bruce Springsteen – "I'm on Fire" Bruce Springsteen – "Glory Days" Awards/nominations Films Awards for Honeydripper: Outstanding Independent or Foreign Film (Win) – 2008 NAACP Image Award Outstanding Writing in a Motion Picture (Theatrical or Television) (Nominated) – John Sayles – 2008 NAACP Image Awards Top 10 Independent Films of 2007 – National Board of Review of Motion Pictures Best Screenplay (Win) – John Sayles – 2007 San Sebastián International Film Festival (Tied with Gracia Querejeta and David Planell for Siete mesas de billar francés (2007)) Award for Silver City: Golden Seashell Award for Best Film (Nominated) – John Sayles – 2004 San Sebastián International Film Festival Awards for Sunshine State: Golden Orange Award (Win) – John Sayles – 2002 Florida Film Critics Circle Awards Special Mention For Excellence In Filmmaking (Win) – 2002 National Board of Review Awards for Limbo: Best Director Golden Space Needle Award (Win) – John Sayles −1999 Seattle International Film Festival Outstanding Indies (Win) – 1999 National Board of Review Awards for Men with Guns/Hombres armados: Best Foreign Independent Film (Nominated) – 1998 British Independent Film Awards Best Foreign Film (Nominated) – 1999 Golden Globes Peace Award (Nominated) – 1998 Political Film Society FIPRESCI Prize (Win) – John Sayles – 1997 San Sebastián International Film Festival OCIC Award (Win) – John Sayles – 1997 San Sebastián International Film Festival Solidarity Award (Win) – John Sayles – 1997 San Sebastián International Film Festival Golden Seashell Award for Best Film (Nominated) – John Sayles – 1997 San Sebastián International Film Festival Awards for Lone Star: Best Original Screenplay (Nominated) – John Sayles – 1997 Academy Awards Best Original Screenplay (Nominated) – John Sayles – 1997 BAFTA Awards Best Screenplay, Motion Picture (Nominated) – John Sayles – 1997 Golden Globes Best Screenplay Written Directly for the Screen (Nominated) – John Sayles – 1997 Writers Guild of America Award Best Picture (Nominated) – 1997 Broadcast Film Critics Association Awards Best Motion Picture Original Screenplay (Win) – John Sayles – 1997 Golden Satellite Awards Best Motion Picture – Drama (Nominated) – Maggie Renzi & R. Paul Miller – 1997 Golden Satellite Awards Best Screenplay (Nominated) – John Sayles – 1997 Independent Spirit Awards Best Film (Win) – Lone Star – 1996 Lone Star Film & Television Awards Best Director (Win) – John Sayles – 1996 Lone Star Film & Television Awards Best Screenplay (Win) – John Sayles – 1996 Lone Star Film & Television Awards Special Achievement Award for Outstanding Feature Film (Win) – 1996 NCLR Bravo Awards Best Director (Win) – John Sayles – 1997 Southeastern Film Critics Association Awards Awards for The Secret of Roan Inish: Best Genre Video Release (Nominated) – 1996 Academy of Science Fiction, Fantasy & Horror Films International Critics Award (Win) – John Sayles – 1996 Gérardmer Film Festival Best Director (Nominated) – John Sayles – 1996 Independent Spirit Awards Best Screenplay (Nominated) – John Sayles – 1996 Independent Spirit Awards Awards for Passion Fish: Best Original Screenplay (Nominated) – John Sayles – 1993 Academy Awards Golden Spur Award (Win) – John Sayles – 1993 Flanders International Film Festival Best Screenplay Written Directly for the Screen (Nominated) – John Sayles – 1993 Writers Guild of America Awards for City of Hope: Critics Award (Nominated) – John Sayles – 1991 Deauville American Film Festival Special Award, Democracy Award (Win) – 1992 Political Film Society Tokyo Grand Prix Award (Win) – John Sayles – 1991 Tokyo International Film Festival Awards for Matewan: Critics Award (Nominated) – John Sayles – 1987 Deauville American Film Festival Best Director (Nominated) – John Sayles – 1988 Independent Spirit Awards Best Screenplay (Nominated) – John Sayles – 1988 Independent Spirit Award Human Rights Award (Win) – 1988 Political Film Society Awards for The Brother from Another Planet: Best Screenplay Caixa de Catalunya Award (Win) – John Sayles – 1984 Catalan International Film Festival, Sitges, Spain Grand Jury Prize – Dramatic (Nominated) – John Sayles – 1985 USA Film Festival (later became the Sundance Film Festival) Awards for Return of the Secaucus 7: Best Independent Film (Win) – 1981 Boston Society of Film Critics Awards Best Screenplay (Win) – John Sayles – 1980 Los Angeles Film Critics Association Awards National Film Registry – 1997 Library of Congress, National Film Preservation Board Best Comedy Written Directly for the Screen (Nominated) – John Sayles – 1981 Writers Guild of America Best Screenplay (Nominated) – John Sayles – 1980 New York Film Critics Circle Second Place – 1981 US Film Festival (later became the Sundance Film Festival) Other recognition Sayles's first published story, "I-80 Nebraska", won an O. Henry Award; his novel, Union Dues, was nominated for a National Book Award as well as the National Book Critics Circle Award. In 1983, Sayles received the John D. MacArthur Award, given to 20 Americans in diverse fields each year for their innovative work. He has also been the recipient of the Eugene V. Debs Award, the John Steinbeck Award and the John Cassavetes Award. He was honored with the Ian McLellan Hunter Award for Lifetime Achievement by the Writers Guild of America (1999). Recurring collaborators Actors who have regularly worked with Sayles include Maggie Renzi, David Strathairn, Joe Morton, Chris Cooper, Mary McDonnell, Vincent Spano, Kevin Tighe, Josh Mostel, Tom Wright, Gordon Clapp and Angela Bassett. See also Night Skies – for a more complete history of how the proposed Close Encounters of the Third Kind sequel became the E.T. the Extra-Terrestrial story Further reading Diane Carson and Heidi Kenaga, eds., Sayles Talk: New Perspectives on Independent Filmmaker John Sayles, Wayne State University Press, 2006 John Sayles, Thinking in Pictures: The Making of the Movie Matewan, Da Capo Press, 2003 References External links Senses of Cinema: Great Directors Critical Database Interview April 2007 by Cathy Pryor in the London Independent The Rumpus interview Return of the Independent: Sayles on Sayles (5-part interview) from siffblog.com 1950 births Actors from Schenectady, New York American male screenwriters American people of Irish descent Edgar Award winners Living people MacArthur Fellows Writers from Schenectady, New York Film directors from New York (state) Screenwriters from New York (state) Williams College alumni
16265
https://en.wikipedia.org/wiki/Jug%20%28disambiguation%29
Jug (disambiguation)
A jug is a type of container commonly used to hold a liquid. Jug may also refer to: Places The Jug, an island in West Virginia, United States Jug Rock, a natural geological formation outside of Shoals, Indiana, United States Jug Sport Hall, an indoor arena in Osijek, Croatia Cepotina, a Serbian Army base also known as Jug Jug II, a city district of Osijek, Croatia Slang Jugs, a slang term for women's breasts, especially large ones Jug, a slang term for prison Australian term for an electric kettle, New Zealand term for a kettle People Jug (nickname) Jug Suraiya, Indian journalist, author and columnist Jug (surname) Other uses Jug, a crossbreed between a Jack Russell terrier and a pug Jug, a nickname for the P-47 Thunderbolt fighter aircraft VK Jug, a water polo club from Dubrovnik, Croatia Jug (instrument), used for rhythmic bass accompaniment Jug wine, a term used for inexpensive table wine Jug Tavern, a historic structure in Ossining, New York, United States Jug (album), by jazz saxophonist Gene Ammons Jug fishing, a method of fishing that uses lines suspended from floating jugs See also Yug (disambiguation) Juggs, a U.S. soft-core pornographic magazine Jugging
16397
https://en.wikipedia.org/wiki/Jacques%20Maroger
Jacques Maroger
Jacques Maroger (; 1884–1962) was a painter and the technical director of the Louvre Museum's laboratory in Paris. He devoted his life to understanding the oil-based media of the Old Masters. He emigrated to the United States in 1939 and became an influential teacher. His book, The Secret Formulas and Techniques of the Masters, has been criticized by some modern writers on painting who say that the painting medium Maroger promoted is unsound. Training and early career In 1907, Maroger began to study with Louis Anquetin and worked under his direction until Anquetin's death in 1932. Anquetin worked closely and exhibited with the artists Vincent van Gogh, Charles Angrand, Émile Bernard, Paul Gauguin, Camille Pissarro, Georges Seurat, Paul Signac and Henri de Toulouse-Lautrec. He was very active in the Impressionist movement of the time. In his later years, Anquetin became very interested in the works of the Flemish masters. As Maroger's teacher, Anquetin provided guidance in the study of drawing, anatomy and master painting techniques. Maroger began to become famous around 1931, when the National Academy of Design in New York City reported Maroger's painting discoveries. From 1930 to 1939, Maroger started to work at the Louvre Museum in Paris as Technical Director of the Louvre Laboratory. He served as a professor at the Louvre School, a Member of the Conservation Committee, General Secretary of the International Experts, and President of the Restorers of France. In 1937, he received the Légion d'honneur, and his pride at the honor is reflected in his self-portrait of the time, in which one can see his Legion pin on his lapel. He emigrated to the United States in 1939 and became a lecturer at the Parsons School of Design in New York. His New York students, Reginald Marsh, John Koch, Fairfield Porter and Frank Mason adopted his Old Master painting techniques, and taught it in turn to their own students. In 1942, Maroger became a Professor at the Maryland Institute College of Art in Baltimore and established a school of painting. At the Maryland Institute he led a group of painters who came to be known as the Baltimore Realists, including the painters Earl Hofmann, Thomas Rowe, Joseph Sheppard, Ann Didusch Schuler, Frank Redelius, John Bannon, Evan Keehn, and Melvin Miller. Maroger published The Secret Formulas and Techniques of the Masters in 1948. When Maroger's book became available, Reginald Marsh drew on Maroger's book-jacket an airplane dropping an atomic bomb on the Maryland Art Institute, a reference to the controversy Maroger was causing in the local press over the abstract art versus realism debate. Maroger's formula and techniques have been studied by many modern painters who wish to obtain the paint quality of the Old Masters. The "secret formula" that Maroger devised during his lifetime included the main ingredient white lead. White lead when cooked into linseed oil acts as a drying agent, accelerating the polymerization of the oil film. Maroger claimed to have introduced to the modern day artist what the masters achieved centuries before in their paintings, a way to ensure permanence and color quality in oils without sacrificing fluid and subtle paint handling. Equipped with these formulas, the artist could once again blend his paint easily without losing control of his brush. The paint stays where it is applied and does not run off the panel. It dries very fast so that he can paint on the same areas the very next day, which speeds up painting. Frank Redelius, one of Maroger's protégés from the Baltimore Realists group, wrote a book that updates, builds upon and revises Jacques Maroger's research of the painting techniques and formulas of the Old Masters. Redelius was assisting Maroger with a revision of The Secret Formulas and Techniques of the Masters before Maroger's death in 1962. Frank Redelius' book, published in 2009, is titled The Master Keys: A Painter's Treatise On The Pictorial Technique Of Oil Painting. Critics of Maroger Maroger has been criticized by some modern writers on painting because of his bold claims about having found the secret formulas of the Masters. The current proprietary Maroger's Medium is in fact the jelly-like medium also known as Megilp (Macgilp, McGuilp, etc.). This material, made by mixing heavy mastic varnish with a linseed oil that has been cooked to blackness with litharge or white lead, was introduced in the late 18th century and employed extensively during the 19th [and therefore is not, as claimed by some, of centuries old pedigree]. The archival quality of the medium itself is controversial in art circles, in part because its documented use dates back less than a century. This is from Michael Skalka, Conservation Administrator, National Gallery of Art, Washington, DC.: See the work of Lance Mayer and Gay Myers for more information on Curry and Maroger. This criticism can be misleading, however. Many of the media involved in Curry's work (and other followers of Maroger) bear no resemblance whatsoever to the modern mastic varnish/black oil recipe. Maroger medium which is not made properly may contain a large amount of dirt and impurities from improperly filtered mastic varnish, or the black oil may be overcooked, both of which would contribute to darkening and weakening of the work. In addition the overuse of megilp media (or any medium for that matter) tends to create weak paint films. Conservation science has shown that the presence of natural resins like mastic in the paint film causes embrittlement, darkening, and continued solubility. See the work of Leslie Carlyle or Joyce Townsend for problems related to 18th-century painting that contain megilp. Lost old master formulas by Maroger Six formulas of Maroger taken from his book on painting formulas Lead Medium – attributed to Antonello da Messina – One part litharge (yellow lead oxide) or lead white, combined by cooking with three to four parts linseed. Lead Medium – attributed to Leonardo da Vinci – One part litharge or lead white, combined by cooking with three to four parts raw linseed oil, and three to four parts water. Lead Medium – attributed to the Venetian painters – Giorgione, Titian and Tintoretto – One or two parts litharge or lead white, combined by cooking with 20 parts raw linseed or walnut oil. Lead Medium – attributed to Peter Paul Rubens -This medium was allegedly based on the black oil of Giorgione with an addition of mastic resin, Venice turpentine and beeswax. One or two parts litharge or lead white, combines by cooking with 20 parts raw linseed. A little more than one spoonful of "black oil" combined with even one spoonful of mastic varnish resulted in the "jelly" medium thought to be Megilp (another name of Maroger media). Lead Medium – (attributed to the "Little Dutch Masters") This medium was the same as the one used by Rubens, but did not include beeswax. Lead Medium – attributed to Velázquez – One part verdigris (derived from copper – this material is substituted for the lead-based metallic driers), combined by cooking with 20 parts raw linseed or walnut oil. The majority of these recipes are not employed today, as there are few companies that produce them. The primary form of "Maroger medium" known today is black oil ("Giorgione's" medium) and mastic varnish combined in approximately equal parts to form a gel. While Maroger medium is usually mixed directly with oil paints, its proportion should be kept to no more than 20% of the mixture. A useful technique is to rub a very thin film of Maroger medium over the area to be painted and paint into that—known as "painting into the couch." This lubricates the brush stroke. Maroger medium (or any other painting medium, for that matter) should never be used as a final picture varnish, as Maroger requires reaction by admixture with oil paint in order to dry. The reduced availability of lead, combined with injunctions against lead use in household products and other factors has caused most major paint makers to discontinue the production of Maroger's medium. Many paint makers now offer faux-maroger's media or faux-megilps, generally made by substituting different materials, such as lime, for genuine lead, or (as in the case of Gamblin's Neo-Megilp) by creating a similar product out of specially thickened alkyd medium. These products produce effects similar to, but not the same as those of real Maroger medium, which depends on specific chemical reactions between leaded oil, mastic resin, and turpentine (the mastic varnish vehicle). Home and Studio The white gingerbread cottage that was Maroger's home in Baltimore is found on the east campus of Loyola College in Maryland and is used for drawing and painting courses. The building, created in the style of a Parisian studio, is aptly called the Maroger Art Studio. References External links The Secret Formulas and Techniques of the Masters by Jacques Maroger How to Paint Like the Old Masters by Joseph Sheppard The Master Keys: A Painter's Treatise on the Pictorial Technique of Oil Painting by Franklin H. Redelius The Schuler School of Fine Arts, Baltimore, MD, USA http://www.schulerschool.com/ Joan Brady Studios, Annapolis, MD, USA Source for authentic Maroger Medium, Black Oil, and Mastic Varnish https://web.archive.org/web/20140320024810/http://bradystudios.net/ 1884 births 1962 deaths 20th-century French painters 20th-century American male artists French male painters 20th-century American painters American male painters Maryland Institute College of Art faculty Recipients of the Legion of Honour French emigrants to the United States
16550
https://en.wikipedia.org/wiki/John%2C%20King%20of%20England
John, King of England
John (24 December 1166 – 19 October 1216) was King of England from 1199 until his death. He lost the Duchy of Normandy and most of his other French lands to King Philip II of France, resulting in the collapse of the Angevin Empire and contributing to the subsequent growth in power of the French Capetian dynasty during the 13th century. The baronial revolt at the end of John's reign led to the sealing of , a document considered an early step in the evolution of the constitution of the United Kingdom. John was the youngest of the four surviving sons of King Henry II of England and Duchess Eleanor of Aquitaine. He was nicknamed John Lackland (Norman French: Jean sans Terre ) because he was not expected to inherit significant lands. He became Henry's favourite child following the failed revolt of 1173–1174 by his brothers Henry the Young King, Richard, and Geoffrey against the King. John was appointed Lord of Ireland in 1177 and given lands in England and on the continent. He unsuccessfully attempted a rebellion against the royal administrators of his brother, King Richard, while Richard was participating in the Third Crusade, but he was proclaimed king after Richard died in 1199. He came to an agreement with Philip II of France to recognise John's possession of the continental Angevin lands at the peace treaty of Le Goulet in 1200. When war with France broke out again in 1202, John achieved early victories, but shortages of military resources and his treatment of Norman, Breton, and Anjou nobles resulted in the collapse of his empire in northern France in 1204. He spent much of the next decade attempting to regain these lands, raising huge revenues, reforming his armed forces and rebuilding continental alliances. His judicial reforms had a lasting effect on the English common law system, as well as providing an additional source of revenue. An argument with Pope Innocent III led to John's excommunication in 1209, a dispute he finally settled in 1213. John's attempt to defeat Philip in 1214 failed because of the French victory over John's allies at the battle of Bouvines. When he returned to England, John faced a rebellion by many of his barons, who were unhappy with his fiscal policies and his treatment of many of England's most powerful nobles. Although John and the barons agreed to the peace treaty in 1215, neither side complied with its conditions. Civil war broke out shortly afterwards, with the barons aided by Louis VIII of France. It soon descended into a stalemate. John died of dysentery contracted while on campaign in eastern England during late 1216; supporters of his son Henry III went on to achieve victory over Louis and the rebel barons the following year. Contemporary chroniclers were mostly critical of John's performance as king, and his reign has since been the subject of significant debate and periodic revision by historians from the 16th century onwards. Historian Jim Bradbury has summarised the current historical opinion of John's positive qualities, observing that John is today usually considered a "hard-working administrator, an able man, an able general". Nonetheless, modern historians agree that he also had many faults as king, including what historian Ralph Turner describes as "distasteful, even dangerous personality traits", such as pettiness, spitefulness, and cruelty. These negative qualities provided extensive material for fiction writers in the Victorian era, and John remains a recurring character within Western popular culture, primarily as a villain in films and stories depicting the Robin Hood legends. Early life (1166–1189) Childhood and the Angevin inheritance John was born on 24 December 1166. His father, King Henry II of England, had inherited significant territories along the Atlantic seaboardAnjou, Normandy and Englandand expanded his empire by conquering Brittany. John's mother was Eleanor, the powerful duchess of Aquitaine, who had a tenuous claim to Toulouse and Auvergne in southern France and was the former wife of King Louis VII of France. The territories of Henry and Eleanor formed the Angevin Empire, named after Henry's paternal title as Count of Anjou and, more specifically, its seat in Angers. The Empire, however, was inherently fragile: although all the lands owed allegiance to Henry, the disparate parts each had their own histories, traditions and governance structures. As one moved south through Anjou and Aquitaine, the extent of Henry's power in the provinces diminished considerably, scarcely resembling the modern concept of an empire at all. Some of the traditional ties between parts of the empire such as Normandy and England were slowly dissolving over time. The future of the empire upon Henry's eventual death was not secure: although the custom of primogeniture, under which an eldest son would inherit all his father's lands, was slowly becoming more widespread across Europe, it was less popular amongst the Norman kings of England. Most believed that Henry would divide the empire, giving each son a substantial portion, and hoping that his children would continue to work together as allies after his death. To complicate matters, much of the Angevin empire was held by Henry only as a vassal of the king of France of the rival line of the House of Capet. Henry had often allied himself with the Holy Roman Emperor against France, making the feudal relationship even more challenging. Shortly after his birth, John was passed from Eleanor into the care of a wet nurse, a traditional practice for medieval noble families. Eleanor then left for Poitiers, the capital of Aquitaine, and sent John and his sister Joan north to Fontevrault Abbey. This may have been done with the aim of steering her youngest son, with no obvious inheritance, towards a future ecclesiastical career. Eleanor spent the next few years conspiring against Henry and neither parent played a part in John's very early life. John was probably, like his brothers, assigned a magister whilst he was at Fontevrault, a teacher charged with his early education and with managing the servants of his immediate household; John was later taught by Ranulf de Glanvill, a leading English administrator. John spent some time as a member of the household of his eldest living brother Henry the Young King, where he probably received instruction in hunting and military skills. John grew up to be around tall, relatively short, with a "powerful, barrel-chested body" and dark red hair; he looked to contemporaries like an inhabitant of Poitou. John enjoyed reading and, unusually for the period, built up a travelling library of books. He enjoyed gambling, in particular at backgammon, and was an enthusiastic hunter, even by medieval standards. He liked music, although not songs. John would become a "connoisseur of jewels", building up a large collection, and became famous for his opulent clothes and also, according to French chroniclers, for his fondness for bad wine. As John grew up, he became known for sometimes being "genial, witty, generous and hospitable"; at other moments, he could be jealous, over-sensitive and prone to fits of rage, "biting and gnawing his fingers" in anger. Early life During John's early years, Henry attempted to resolve the question of his succession. Henry the Young King had been crowned King of England in 1170, but was not given any formal powers by his father; he was also promised Normandy and Anjou as part of his future inheritance. His brother Richard was to be appointed the count of Poitou with control of Aquitaine, whilst his brother Geoffrey was to become the duke of Brittany. At this time it seemed unlikely that John would ever inherit substantial lands, and he was jokingly nicknamed "Lackland" by his father. Henry II wanted to secure the southern borders of Aquitaine and decided to betroth his youngest son to Alais, the daughter and heiress of Humbert III of Savoy. As part of this agreement John was promised the future inheritance of Savoy, Piedmont, Maurienne, and the other possessions of Count Humbert. For his part in the potential marriage alliance, Henry II transferred the castles of Chinon, Loudun and Mirebeau into John's name; as John was only five years old his father would continue to control them for practical purposes. Henry the Young King was unimpressed by this; although he had yet to be granted control of any castles in his new kingdom, these were effectively his future property and had been given away without consultation. Alais made the trip over the Alps and joined Henry II's court, but she died before marrying John, which left the prince once again without an inheritance. In 1173 John's elder brothers, backed by Eleanor, rose in revolt against Henry in the short-lived rebellion of 1173 to 1174. Growing irritated with his subordinate position to Henry II and increasingly worried that John might be given additional lands and castles at his expense, Henry the Young King travelled to Paris and allied himself with Louis VII. Eleanor, irritated by her husband's persistent interference in Aquitaine, encouraged Richard and Geoffrey to join their brother Henry in Paris. Henry II triumphed over the coalition of his sons, but was generous to them in the peace settlement agreed at Montlouis. Henry the Young King was allowed to travel widely in Europe with his own household of knights, Richard was given Aquitaine back, and Geoffrey was allowed to return to Brittany; only Eleanor was imprisoned for her role in the revolt. John had spent the conflict travelling alongside his father, and was given widespread possessions across the Angevin empire as part of the Montlouis settlement; from then onwards, most observers regarded John as Henry II's favourite child, although he was the furthest removed in terms of the royal succession. Henry II began to find more lands for John, mostly at various nobles' expense. In 1175 he appropriated the estates of the late Earl of Cornwall and gave them to John. The following year, Henry disinherited the sisters of Isabella of Gloucester, contrary to legal custom, and betrothed John to the now extremely wealthy Isabella. In 1177, at the Council of Oxford, Henry dismissed William FitzAldelm as the Lord of Ireland and replaced him with the ten-year-old John. Henry the Young King fought a short war with his brother Richard in 1183 over the status of England, Normandy and Aquitaine. Henry II moved in support of Richard, and Henry the Young King died from dysentery at the end of the campaign. With his primary heir dead, Henry rearranged the plans for the succession: Richard was to be made King of England, albeit without any actual power until the death of his father; Geoffrey would retain Brittany; and John would now become the Duke of Aquitaine in place of Richard. Richard refused to give up Aquitaine; Henry II was furious and ordered John, with help from Geoffrey, to march south and retake the duchy by force. The two attacked the capital of Poitiers, and Richard responded by attacking Brittany. The war ended in stalemate and a tense family reconciliation in England at the end of 1184. In 1185 John made his first visit to Ireland, accompanied by 300 knights and a team of administrators. Henry had tried to have John officially proclaimed King of Ireland, but Pope Lucius III would not agree. John's first period of rule in Ireland was not a success. Ireland had only recently been conquered by Anglo-Norman forces, and tensions were still rife between Henry II, the new settlers and the existing inhabitants. John infamously offended the local Irish rulers by making fun of their unfashionable long beards, failed to make allies amongst the Anglo-Norman settlers, began to lose ground militarily against the Irish and finally returned to England later in the year, blaming the viceroy, Hugh de Lacy, for the fiasco. The problems amongst John's wider family continued to grow. His elder brother Geoffrey died during a tournament in 1186, leaving a posthumous son, Arthur, and an elder daughter, Eleanor. Geoffrey's death brought John slightly closer to the throne of England. The uncertainty about what would happen after Henry's death continued to grow; Richard was keen to join a new crusade and remained concerned that whilst he was away Henry would appoint John his formal successor. Richard began discussions about a potential alliance with Philip II in Paris during 1187, and the next year Richard gave homage to Philip in exchange for support for a war against Henry. Richard and Philip fought a joint campaign against Henry, and by the summer of 1189 the king made peace, promising Richard the succession. John initially remained loyal to his father, but changed sides once it appeared that Richard would win. Henry died shortly afterwards. Richard's reign (1189–1199) When Richard became king in September 1189, he had already declared his intention of joining the Third Crusade. He set about raising the huge sums of money required for this expedition through the sale of lands, titles and appointments, and attempted to ensure that he would not face a revolt while away from his empire. John was made Count of Mortain, was married to the wealthy Isabella of Gloucester, and was given valuable lands in Lancaster and the counties of Cornwall, Derby, Devon, Dorset, Nottingham and Somerset, all with the aim of buying his loyalty to Richard whilst the King was on crusade. Richard retained royal control of key castles in these counties, thereby preventing John from accumulating too much military and political power. The King named his four-year-old nephew Arthur as his heir. In return, John promised not to visit England for the next three years, thereby in theory giving Richard adequate time to conduct a successful crusade and return from the Levant without fear of John seizing power. Richard left political authority in England—the post of justiciar—jointly in the hands of Bishop Hugh de Puiset and William de Mandeville, 3rd Earl of Essex, and made William Longchamp, the Bishop of Ely, his chancellor. Mandeville immediately died, and Longchamp took over as joint justiciar with Puiset, which would prove a less than satisfactory partnership. Eleanor, the queen mother, convinced Richard to allow John into England in his absence. The political situation in England rapidly began to deteriorate. Longchamp refused to work with Puiset and became unpopular with the English nobility and clergy. John exploited this unpopularity to set himself up as an alternative ruler with his own royal court, complete with his own justiciar, chancellor and other royal posts, and was happy to be portrayed as an alternative regent, and possibly the next king. Armed conflict broke out between John and Longchamp, and by October 1191 Longchamp was isolated in the Tower of London with John in control of the city of London, thanks to promises John had made to the citizens in return for recognition as Richard's heir presumptive. At this point Walter of Coutances, the Archbishop of Rouen, returned to England, having been sent by Richard to restore order. John's position was undermined by Walter's relative popularity and by the news that Richard had married whilst in Cyprus, which presented the possibility that Richard would have legitimate children and heirs. The political turmoil continued. John began to explore an alliance with King Philip II of France, who had returned from the crusade in late 1191. John hoped to acquire Normandy, Anjou and the other lands in France held by Richard in exchange for allying himself with Philip. John was persuaded not to pursue an alliance by his mother. Longchamp, who had left England after Walter's intervention, now returned, and argued that he had been wrongly removed as justiciar. John intervened, suppressing Longchamp's claims in return for promises of support from the royal administration, including a reaffirmation of his position as heir to the throne. When Richard still did not return from the crusade, John began to assert that his brother was dead or otherwise permanently lost. Richard had in fact been captured shortly before Christmas 1192, while en route to England, by Duke Leopold V of Austria and was handed over to Emperor Henry VI, who held him for ransom. John seized the opportunity and went to Paris, where he formed an alliance with Philip. He agreed to set aside his wife, Isabella of Gloucester, and marry Philip's sister, Alys, in exchange for Philip's support. Fighting broke out in England between forces loyal to Richard and those being gathered by John. John's military position was weak and he agreed to a truce; in early 1194 the King finally returned to England, and John's remaining forces surrendered. John retreated to Normandy, where Richard finally found him later that year. Richard declared that John—despite being 27 years old—was merely "a child who has had evil counsellors" and forgave him, but removed his lands with the exception of Ireland. For the remaining years of Richard's reign, John supported his brother on the continent, apparently loyally. Richard's policy on the continent was to attempt to regain through steady, limited campaigns the castles he had lost to Philip II whilst on crusade. He allied himself with the leaders of Flanders, Boulogne and the Holy Roman Empire to apply pressure on Philip from Germany. In 1195 John successfully conducted a sudden attack and siege of Évreux castle, and subsequently managed the defences of Normandy against Philip. The following year, John seized the town of Gamaches and led a raiding party within of Paris, capturing the Bishop of Beauvais. In return for this service, Richard withdrew his (ill-will) towards John, restored him to the county of Gloucestershire and made him again the Count of Mortain. Early reign (1199–1204) Accession to the throne, 1199 After Richard's death on 6 April 1199 there were two potential claimants to the Angevin throne: John, whose claim rested on being the sole surviving son of Henry II, and young Arthur I of Brittany, who held a claim as the son of John's elder brother Geoffrey. Richard appears to have started to recognise John as his heir presumptive in the final years before his death, but the matter was not clear-cut and medieval law gave little guidance as to how the competing claims should be decided. With Norman law favouring John as the only surviving son of Henry II and Angevin law favouring Arthur as the only son of Henry's elder son, the matter rapidly became an open conflict. John was supported by the bulk of the English and Norman nobility and was crowned at Westminster Abbey, backed by his mother, Eleanor. Arthur was supported by the majority of the Breton, Maine and Anjou nobles and received the support of Philip II, who remained committed to breaking up the Angevin territories on the continent. With Arthur's army pressing up the Loire Valley towards Angers and Philip's forces moving down the valley towards Tours, John's continental empire was in danger of being cut in two. Warfare in Normandy at the time was shaped by the defensive potential of castles and the increasing costs of conducting campaigns. The Norman frontiers had limited natural defences but were heavily reinforced with castles, such as Château Gaillard, at strategic points, built and maintained at considerable expense. It was difficult for a commander to advance far into fresh territory without having secured his lines of communication by capturing these fortifications, which slowed the progress of any attack. Armies of the period could be formed from either feudal or mercenary forces. Feudal levies could be raised only for a fixed length of time before they returned home, forcing an end to a campaign; mercenary forces, often called Brabançons after the Duchy of Brabant but actually recruited from across northern Europe, could operate all year long and provide a commander with more strategic options to pursue a campaign, but cost much more than equivalent feudal forces. As a result, commanders of the period were increasingly drawing on larger numbers of mercenaries. After his coronation, John moved south into France with military forces and adopted a defensive posture along the eastern and southern Normandy borders. Both sides paused for desultory negotiations before the war recommenced; John's position was now stronger, thanks to confirmation that the counts Baldwin IX of Flanders and Renaud of Boulogne had renewed the anti-French alliances they had previously agreed to with Richard. The powerful Anjou nobleman William des Roches was persuaded to switch sides from Arthur to John; suddenly the balance seemed to be tipping away from Philip and Arthur in favour of John. Neither side was keen to continue the conflict, and following a papal truce the two leaders met in January 1200 to negotiate possible terms for peace. From John's perspective, what then followed represented an opportunity to stabilise control over his continental possessions and produce a lasting peace with Philip in Paris. John and Philip negotiated the May 1200 Treaty of Le Goulet; by this treaty, Philip recognised John as the rightful heir to Richard in respect to his French possessions, temporarily abandoning the wider claims of his client, Arthur. John, in turn, abandoned Richard's former policy of containing Philip through alliances with Flanders and Boulogne, and accepted Philip's right as the legitimate feudal overlord of John's lands in France. John's policy earned him the disrespectful title of "John Softsword" from some English chroniclers, who contrasted his behaviour with his more aggressive brother, Richard. Second marriage and consequences, 1200–1202 The new peace would last only two years; war recommenced in the aftermath of John's decision in August 1200 to marry Isabella of Angoulême. In order to remarry, John first needed to abandon his wife Isabella, Countess of Gloucester; the King accomplished this by arguing that he had failed to get the necessary papal dispensation to marry the Countess in the first place—as a cousin, John could not have legally wedded her without this. It remains unclear why John chose to marry Isabella of Angoulême. Contemporary chroniclers argued that John had fallen deeply in love with her, and John may have been motivated by desire for an apparently beautiful, if rather young, girl (Isabella of Angoulême was either 12 or 14 at the time of their marriage). On the other hand, the Angoumois lands that came with her were strategically vital to John: by marrying Isabella, John was acquiring a key land route between Poitou and Gascony, which significantly strengthened his grip on Aquitaine. Isabella, however, was already engaged to Hugh IX of Lusignan, an important member of a key Poitou noble family and brother of Raoul I, Count of Eu, who possessed lands along the sensitive eastern Normandy border. Just as John stood to benefit strategically from marrying Isabella, so the marriage threatened the interests of the Lusignans, whose own lands currently provided the key route for royal goods and troops across Aquitaine. Rather than negotiating some form of compensation, John treated Hugh "with contempt"; this resulted in a Lusignan uprising that was promptly crushed by John, who also intervened to suppress Raoul in Normandy. Although John was the Count of Poitou and therefore the rightful feudal lord over the Lusignans, they could legitimately appeal John's actions in France to his own feudal lord, Philip. Hugh did exactly this in 1201 and Philip summoned John to attend court in Paris in 1202, citing the Le Goulet treaty to strengthen his case. John was unwilling to weaken his authority in western France in this way. He argued that he need not attend Philip's court because of his special status as the Duke of Normandy, who was exempt by feudal tradition from being called to the French court. Philip argued that he was summoning John not as the Duke of Normandy, but as the Count of Poitou, which carried no such special status. When John still refused to come, Philip declared John in breach of his feudal responsibilities, reassigned all of John's lands that fell under the French crown to Arthur—with the exception of Normandy, which he took back for himself—and began a fresh war against John. Loss of Normandy, 1202–1204 John initially adopted a defensive posture similar to that of 1199: avoiding open battle and carefully defending his key castles. John's operations became more chaotic as the campaign progressed, and Philip began to make steady progress in the east. John became aware in July that Arthur's forces were threatening his mother, Eleanor, at Mirebeau Castle. Accompanied by William de Roches, his seneschal in Anjou, he swung his mercenary army rapidly south to protect her. His forces caught Arthur by surprise and captured the entire rebel leadership at the battle of Mirebeau. With his southern flank weakening, Philip was forced to withdraw in the east and turn south himself to contain John's army. John's position in France was considerably strengthened by the victory at Mirebeau, but John's treatment of his new prisoners and of his ally, William de Roches, quickly undermined these gains. De Roches was a powerful Anjou noble, but John largely ignored him, causing considerable offence, whilst the King kept the rebel leaders in such bad conditions that twenty-two of them died. At this time most of the regional nobility were closely linked through kinship, and this behaviour towards their relatives was regarded as unacceptable. William de Roches and other of John's regional allies in Anjou and Brittany deserted him in favour of Philip, and Brittany rose in fresh revolt. John's financial situation was tenuous: once factors such as the comparative military costs of materiel and soldiers were taken into account, Philip enjoyed a considerable, although not overwhelming, advantage of resources over John. Further desertions of John's local allies at the beginning of 1203 steadily reduced his freedom to manoeuvre in the region. He attempted to convince Pope Innocent III to intervene in the conflict, but Innocent's efforts were unsuccessful. As the situation became worse for John, he appears to have decided to have Arthur killed, with the aim of removing his potential rival and of undermining the rebel movement in Brittany. Arthur had initially been imprisoned at Falaise and was then moved to Rouen. After this, Arthur's fate remains uncertain, but modern historians believe he was murdered by John. The annals of Margam Abbey suggest that "John had captured Arthur and kept him alive in prison for some time in the castle of Rouen ... when John was drunk he slew Arthur with his own hand and tying a heavy stone to the body cast it into the Seine." Rumours of the manner of Arthur's death further reduced support for John across the region. Arthur's sister, Eleanor, who had also been captured at Mirebeau, was kept imprisoned by John for many years, albeit in relatively good conditions. In late 1203, John attempted to relieve Château Gaillard, which although besieged by Philip was guarding the eastern flank of Normandy. John attempted a synchronised operation involving land-based and water-borne forces, considered by most historians today to have been imaginative in conception, but overly complex for forces of the period to have carried out successfully. John's relief operation was blocked by Philip's forces, and John turned back to Brittany in an attempt to draw Philip away from eastern Normandy. John successfully devastated much of Brittany, but did not deflect Philip's main thrust into the east of Normandy. Opinions vary amongst historians as to the military skill shown by John during this campaign, with most recent historians arguing that his performance was passable, although not impressive. John's situation began to deteriorate rapidly. The eastern border region of Normandy had been extensively cultivated by Philip and his predecessors for several years, whilst Angevin authority in the south had been undermined by Richard's giving away of various key castles some years before. His use of mercenaries in the central regions had rapidly eaten away his remaining support in this area too, which set the stage for a sudden collapse of Angevin power. John retreated back across the Channel in December, sending orders for the establishment of a fresh defensive line to the west of Chateau Gaillard. In March 1204, Gaillard fell. John's mother Eleanor died the following month. This was not just a personal blow for John, but threatened to unravel the widespread Angevin alliances across the far south of France. Philip moved south around the new defensive line and struck upwards at the heart of the Duchy, now facing little resistance. By August, Philip had taken Normandy and advanced south to occupy Anjou and Poitou as well. John's only remaining possession on the Continent was now the Duchy of Aquitaine. John as king Kingship and royal administration The nature of government under the Angevin monarchs was ill-defined and uncertain. John's predecessors had ruled using the principle of ("force and will"), taking executive and sometimes arbitrary decisions, often justified on the basis that a king was above the law. Both Henry II and Richard had argued that kings possessed a quality of "divine majesty"; John continued this trend and claimed an "almost imperial status" for himself as ruler. During the 12th century, there were contrary opinions expressed about the nature of kingship, and many contemporary writers believed that monarchs should rule in accordance with the custom and the law, and take counsel of the leading members of the realm. There was as yet no model for what should happen if a king refused to do so. Despite his claim to unique authority within England, John would sometimes justify his actions on the basis that he had taken council with the barons. Modern historians remain divided as to whether John suffered from a case of "royal schizophrenia" in his approach to government, or if his actions merely reflected the complex model of Angevin kingship in the early 13th century. John inherited a sophisticated system of administration in England, with a range of royal agents answering to the Royal Household: the Chancery kept written records and communications; the Treasury and the Exchequer dealt with income and expenditure respectively; and various judges were deployed to deliver justice around the kingdom. Thanks to the efforts of men like Hubert Walter, this trend towards improved record keeping continued into his reign. Like previous kings, John managed a peripatetic court that travelled around the kingdom, dealing with both local and national matters as he went. John was very active in the administration of England and was involved in every aspect of government. In part he was following in the tradition of Henry I and Henry II, but by the 13th century the volume of administrative work had greatly increased, which put much more pressure on a king who wished to rule in this style. John was in England for much longer periods than his predecessors, which made his rule more personal than that of previous kings, particularly in previously ignored areas such as the north. The administration of justice was of particular importance to John. Several new processes had been introduced to English law under Henry II, including novel disseisin and mort d'ancestor. These processes meant the royal courts had a more significant role in local law cases, which had previously been dealt with only by regional or local lords. John increased the professionalism of local sergeants and bailiffs, and extended the system of coroners first introduced by Hubert Walter in 1194, creating a new class of borough coroners. The King worked extremely hard to ensure that this system operated well, through judges he had appointed, by fostering legal specialists and expertise, and by intervening in cases himself. He continued to try relatively minor cases, even during military crises. Viewed positively, Lewis Warren considers that John discharged "his royal duty of providing justice ... with a zeal and a tirelessness to which the English common law is greatly endebted". Seen more critically, John may have been motivated by the potential of the royal legal process to raise fees, rather than a desire to deliver simple justice; his legal system also applied only to free men, rather than to all of the population. Nonetheless, these changes were popular with many free tenants, who acquired a more reliable legal system that could bypass the barons, against whom such cases were often brought. John's reforms were less popular with the barons themselves, especially as they remained subject to arbitrary and frequently vindictive royal justice. Economy One of John's principal challenges was acquiring the large sums of money needed for his proposed campaigns to reclaim Normandy. The Angevin kings had three main sources of income available to them, namely revenue from their personal lands, or demesne; money raised through their rights as a feudal lord; and revenue from taxation. Revenue from the royal demesne was inflexible and had been diminishing slowly since the Norman conquest. Matters were not helped by Richard's sale of many royal properties in 1189, and taxation played a much smaller role in royal income than in later centuries. English kings had widespread feudal rights which could be used to generate income, including the scutage system, in which feudal military service was avoided by a cash payment to the King. He derived income from fines, court fees and the sale of charters and other privileges. John intensified his efforts to maximise all possible sources of income, to the extent that he has been described as "avaricious, miserly, extortionate and moneyminded". He also used revenue generation as a way of exerting political control over the barons: debts owed to the crown by the King's favoured supporters might be forgiven; collection of those owed by enemies was more stringently enforced. The result was a sequence of innovative but unpopular financial measures. John levied scutage payments eleven times in his seventeen years as king, as compared to eleven times in total during the reign of the preceding three monarchs. In many cases these were levied in the absence of any actual military campaign, which ran counter to the original idea that scutage was an alternative to actual military service. John maximised his right to demand relief payments when estates and castles were inherited, sometimes charging enormous sums, beyond barons' abilities to pay. Building on the successful sale of sheriff appointments in 1194, the King initiated a new round of appointments, with the new incumbents making back their investment through increased fines and penalties, particularly in the forests. Another innovation of Richard's, increased charges levied on widows who wished to remain single, was expanded under John. John continued to sell charters for new towns, including the planned town of Liverpool, and charters were sold for markets across the kingdom and in Gascony. The King introduced new taxes and extended existing ones. The Jews, who held a vulnerable position in medieval England, protected only by the King, were subject to huge taxes; £44,000 was extracted from the community by the tallage of 1210; much of it was passed on to the Christian debtors of Jewish moneylenders. John created a new tax on income and movable goods in 1207—effectively a version of a modern income tax—that produced £60,000; he created a new set of import and export duties payable directly to the Crown. He found that these measures enabled him to raise further resources through the confiscation of the lands of barons who could not pay or refused to pay. At the start of John's reign there was a sudden change in prices, as bad harvests and high demand for food resulted in much higher prices for grain and animals. This inflationary pressure was to continue for the rest of the 13th century and had long-term economic consequences for England. The resulting social pressures were complicated by bursts of deflation that resulted from John's military campaigns. It was usual at the time for the King to collect taxes in silver, which was then re-minted into new coins; these coins would then be put in barrels and sent to royal castles around the country, to be used to hire mercenaries or to meet other costs. At those times when John was preparing for campaigns in Normandy, for example, huge quantities of silver had to be withdrawn from the economy and stored for months, which unintentionally resulted in periods during which silver coins were simply hard to come by, commercial credit difficult to acquire and deflationary pressure placed on the economy. The result was political unrest across the country. John attempted to address some of the problems with the English currency in 1204 and 1205 by carrying out a radical overhaul of the coinage, improving its quality and consistency. Royal household and John's royal household was based around several groups of followers. One group was the , his immediate friends and knights who travelled around the country with him. They also played an important role in organising and leading military campaigns. Another section of royal followers were the ; these were the senior officials and agents of the King and were essential to his day-to-day rule. Being a member of these inner circles brought huge advantages, as it was easier to gain favours from the King, file lawsuits, marry a wealthy heiress or have one's debts remitted. By the time of Henry II, these posts were increasingly being filled by "new men" from outside the normal ranks of the barons. This intensified under John's rule, with many lesser nobles arriving from the continent to take up positions at court; many were mercenary leaders from Poitou. These men included soldiers who would become infamous in England for their uncivilised behaviour, including Falkes de Breauté, Geard d'Athies, Engelard de Cigongé, and Philip Marc. Many barons perceived the King's household as what Ralph Turner has characterised as a "narrow clique enjoying royal favour at barons' expense" staffed by men of lesser status. This trend for the King to rely on his own men at the expense of the barons was exacerbated by the tradition of Angevin royal ("anger and ill-will") and John's own personality. From Henry II onwards, had come to describe the right of the King to express his anger and displeasure at particular barons or clergy, building on the Norman concept of —royal ill-will. In the Norman period, suffering the King's ill-will meant difficulties in obtaining grants, honours or petitions; Henry II had infamously expressed his fury and ill-will towards Thomas Becket, which ultimately resulted in Becket's death. John now had the additional ability to "cripple his vassals" on a significant scale using his new economic and judicial measures, which made the threat of royal anger all the more serious. John was deeply suspicious of the barons, particularly those with sufficient power and wealth to potentially challenge him. Numerous barons were subjected to his , even including the famous knight William Marshal, 1st Earl of Pembroke, normally held up as a model of utter loyalty. The most infamous case, which went beyond anything considered acceptable at the time, was that of the powerful William de Braose, 4th Lord of Bramber, who held lands in Ireland. De Braose was subjected to punitive demands for money, and when he refused to pay a huge sum of 40,000 marks (equivalent to £26,666 at the time), his wife, Maud, and one of their sons were imprisoned by John, which resulted in their deaths. De Braose died in exile in 1211, and his grandsons remained in prison until 1218. John's suspicions and jealousies meant that he rarely enjoyed good relationships with even the leading loyalist barons. Personal life John's personal life greatly affected his reign. Contemporary chroniclers state that John was sinfully lustful and lacking in piety. It was common for kings and nobles of the period to keep mistresses, but chroniclers complained that John's mistresses were married noblewomen, which was considered unacceptable. John had at least five children with mistresses during his first marriage, and two of those mistresses are known to have been noblewomen. John's behaviour after his second marriage is less clear, however. None of his known illegitimate children were born after he remarried, and there is no actual documentary proof of adultery after that point, although John certainly had female friends amongst the court throughout the period. The specific accusations made against John during the baronial revolts are now generally considered to have been invented for the purposes of justifying the revolt; nonetheless, most of John's contemporaries seem to have held a poor opinion of his sexual behaviour. The character of John's relationship with his second wife, Isabella of Angoulême, is unclear. John married Isabella whilst she was relatively young—her exact date of birth is uncertain, and estimates place her between at most 15 and more probably towards nine years old at the time of her marriage. Even by the standards of the time, she was married whilst very young. John did not provide a great deal of money for his wife's household and did not pass on much of the revenue from her lands, to the extent that historian Nicholas Vincent has described him as being "downright mean" towards Isabella. Vincent concluded that the marriage was not a particularly "amicable" one. Other aspects of their marriage suggest a closer, more positive relationship. Chroniclers recorded that John had a "mad infatuation" with Isabella, and certainly the King and Queen had conjugal relationships between at least 1207 and 1215; they had five children. In contrast to Vincent, historian William Chester Jordan concludes that the pair were a "companionable couple" who had a successful marriage by the standards of the day. John's lack of religious conviction has been noted by contemporary chroniclers and later historians, with some suspecting that he was at best impious, or even atheistic, a very serious issue at the time. Contemporary chroniclers catalogued his various anti-religious habits at length, including his failure to take communion, his blasphemous remarks, and his witty but scandalous jokes about church doctrine, including jokes about the implausibility of the Resurrection of Jesus. They commented on the paucity of John's charitable donations to the Church. Historian Frank McLynn argues that John's early years at Fontevrault, combined with his relatively advanced education, may have turned him against the church. Other historians have been more cautious in interpreting this material, noting that chroniclers also reported his personal interest in the life of St Wulfstan and his friendships with several senior clerics, most especially with Hugh of Lincoln, who was later declared a saint. Financial records show a normal royal household engaged in the usual feasts and pious observances—albeit with many records showing John's offerings to the poor to atone for routinely breaking church rules and guidance. The historian Lewis Warren has argued that the chronicler accounts were subject to considerable bias and the King was "at least conventionally devout", citing his pilgrimages and interest in religious scripture and commentaries. Later reign (1204–1214) Continental policy During the remainder of his reign, John focused on trying to retake Normandy. The available evidence suggests that he did not regard the loss of the Duchy as a permanent shift in Capetian power. Strategically, John faced several challenges: England itself had to be secured against possible French invasion, the sea-routes to Bordeaux needed to be secured following the loss of the land route to Aquitaine, and his remaining possessions in Aquitaine needed to be secured following the death of his mother, Eleanor, in April 1204. John's preferred plan was to use Poitou as a base of operations, advance up the Loire Valley to threaten Paris, pin down the French forces and break Philip's internal lines of communication before landing a maritime force in the Duchy itself. Ideally, this plan would benefit from the opening of a second front on Philip's eastern frontiers with Flanders and Boulogne—effectively a re-creation of Richard's old strategy of applying pressure from Germany. All of this would require a great deal of money and soldiers. John spent much of 1205 securing England against a potential French invasion. As an emergency measure, he recreated a version of Henry II's Assize of Arms of 1181, with each shire creating a structure to mobilise local levies. When the threat of invasion faded, John formed a large military force in England intended for Poitou, and a large fleet with soldiers under his own command intended for Normandy. To achieve this, John reformed the English feudal contribution to his campaigns, creating a more flexible system under which only one knight in ten would actually be mobilised, but would be financially supported by the other nine; knights would serve for an indefinite period. John built up a strong team of engineers for siege warfare and a substantial force of professional crossbowmen. The King was supported by a team of leading barons with military expertise, including William Longespée, 3rd Earl of Salisbury, William the Marshal, Roger de Lacy and, until he fell from favour, the marcher lord William de Braose. John had already begun to improve his Channel forces before the loss of Normandy and he rapidly built up further maritime capabilities after its collapse. Most of these ships were placed along the Cinque Ports, but Portsmouth was also enlarged. By the end of 1204 he had around 50 large galleys available; another 54 vessels were built between 1209 and 1212. William of Wrotham was appointed "keeper of the galleys", effectively John's chief admiral. Wrotham was responsible for fusing John's galleys, the ships of the Cinque Ports and pressed merchant vessels into a single operational fleet. John adopted recent improvements in ship design, including new large transport ships called buisses and removable forecastles for use in combat. Baronial unrest in England prevented the departure of the planned 1205 expedition, and only a smaller force under William Longespée deployed to Poitou. In 1206 John departed for Poitou himself, but was forced to divert south to counter a threat to Gascony from Alfonso VIII of Castile. After a successful campaign against Alfonso, John headed north again, taking the city of Angers. Philip moved south to meet John; the year's campaigning ended in stalemate and a two-year truce was made between the two rulers. During the truce of 1206–1208, John focused on building up his financial and military resources in preparation for another attempt to recapture Normandy. John used some of this money to pay for new alliances on Philip's eastern frontiers, where the growth in Capetian power was beginning to concern France's neighbours. By 1212 John had successfully concluded alliances with his nephew Otto IV, a contender for the crown of Holy Roman Emperor in Germany, as well as with the counts Renaud of Boulogne and Ferdinand of Flanders. The invasion plans for 1212 were postponed because of fresh English baronial unrest about service in Poitou. Philip seized the initiative in 1213, sending his elder son, Louis, to invade Flanders with the intention of next launching an invasion of England. John was forced to postpone his own invasion plans to counter this threat. He launched his new fleet to attack the French at the harbour of Damme. The attack was a success, destroying Philip's vessels and any chances of an invasion of England that year. John hoped to exploit this advantage by invading himself late in 1213, but baronial discontent again delayed his invasion plans until early 1214, in what was his final Continental campaign. Scotland, Ireland and Wales In the late 12th and early 13th centuries the border and political relationship between England and Scotland was disputed, with the kings of Scotland claiming parts of what is now northern England. John's father, Henry II, had forced William the Lion to swear fealty to him at the Treaty of Falaise in 1174. This had been rescinded by Richard I in exchange for financial compensation in 1189, but the relationship remained uneasy. John began his reign by reasserting his sovereignty over the disputed northern counties. He refused William's request for the earldom of Northumbria, but did not intervene in Scotland itself and focused on his continental problems. The two kings maintained a friendly relationship, meeting in 1206 and 1207, until it was rumoured in 1209 that William was intending to ally himself with Philip II of France. John invaded Scotland and forced William to sign the Treaty of Norham, which gave John control of William's daughters and required a payment of £10,000. This effectively crippled William's power north of the border, and by 1212 John had to intervene militarily to support William against his internal rivals. John made no efforts to reinvigorate the Treaty of Falaise, though, and William and his son Alexander II of Scotland in turn remained independent kings, supported by, but not owing fealty to, John. John remained Lord of Ireland throughout his reign. He drew on the country for resources to fight his war with Philip on the continent. Conflict continued in Ireland between the Anglo-Norman settlers and the indigenous Irish chieftains, with John manipulating both groups to expand his wealth and power in the country. During Richard's rule, John had successfully increased the size of his lands in Ireland, and he continued this policy as king. In 1210 the King crossed into Ireland with a large army to crush a rebellion by the Anglo-Norman lords; he reasserted his control of the country and used a new charter to order compliance with English laws and customs in Ireland. John stopped short of trying to actively enforce this charter on the native Irish kingdoms, but historian David Carpenter suspects that he might have done so, had the baronial conflict in England not intervened. Simmering tensions remained with the native Irish leaders even after John left for England. Royal power in Wales was unevenly applied, with the country divided between the marcher lords along the borders, royal territories in Pembrokeshire and the more independent native Welsh lords of North Wales. John took a close interest in Wales and knew the country well, visiting every year between 1204 and 1211 and marrying his illegitimate daughter, Joan, to the Welsh prince Llywelyn the Great. The King used the marcher lords and the native Welsh to increase his own territory and power, striking a sequence of increasingly precise deals backed by royal military power with the Welsh rulers. A major royal expedition to enforce these agreements occurred in 1211, after Llywelyn attempted to exploit the instability caused by the removal of William de Braose, through the Welsh uprising of 1211. John's invasion, striking into the Welsh heartlands, was a military success. Llywelyn came to terms that included an expansion of John's power across much of Wales, albeit only temporarily. Dispute with the Pope and excommunication When the Archbishop of Canterbury, Hubert Walter, died on 13 July 1205, John became involved in a dispute with Pope Innocent III that would lead to the King's excommunication. The Norman and Angevin kings had traditionally exercised a great deal of power over the church within their territories. From the 1040s onwards, however, successive popes had put forward a reforming message that emphasised the importance of the Church being "governed more coherently and more hierarchically from the centre" and established "its own sphere of authority and jurisdiction, separate from and independent of that of the lay ruler", in the words of historian Richard Huscroft. After the 1140s, these principles had been largely accepted within the English Church, albeit with an element of concern about centralising authority in Rome. These changes brought the customary rights of lay rulers such as John over ecclesiastical appointments into question. Pope Innocent was, according to historian Ralph Turner, an "ambitious and aggressive" religious leader, insistent on his rights and responsibilities within the church. John wanted John de Gray, the Bishop of Norwich and one of his own supporters, to be appointed Archbishop of Canterbury, but the cathedral chapter for Canterbury Cathedral claimed the exclusive right to elect the Archbishop. They favoured Reginald, the chapter's sub-prior. To complicate matters, the bishops of the province of Canterbury also claimed the right to appoint the next archbishop. The chapter secretly elected Reginald and he travelled to Rome to be confirmed; the bishops challenged the appointment and the matter was taken before Innocent. John forced the Canterbury chapter to change their support to John de Gray, and a messenger was sent to Rome to inform the papacy of the new decision. Innocent disavowed both Reginald and John de Gray, and instead appointed his own candidate, Stephen Langton. John refused Innocent's request that he consent to Langton's appointment, but the Pope consecrated Langton anyway in June 1207. John was incensed about what he perceived as an abrogation of his customary right as monarch to influence the election. He complained both about the choice of Langton as an individual, as John felt he was overly influenced by the Capetian court in Paris, and about the process as a whole. He barred Langton from entering England and seized the lands of the archbishopric and other papal possessions. Innocent set a commission in place to try to convince John to change his mind, but to no avail. Innocent then placed an interdict on England in March 1208, prohibiting clergy from conducting religious services, with the exception of baptisms for the young, and confessions and absolutions for the dying. John treated the interdict as "the equivalent of a papal declaration of war". He responded by attempting to punish Innocent personally and to drive a wedge between those English clergy that might support him and those allying themselves firmly with the authorities in Rome. John seized the lands of those clergy unwilling to conduct services, as well as those estates linked to Innocent himself; he arrested the illicit concubines that many clerics kept during the period, releasing them only after the payment of fines; he seized the lands of members of the church who had fled England, and he promised protection for those clergy willing to remain loyal to him. In many cases, individual institutions were able to negotiate terms for managing their own properties and keeping the produce of their estates. By 1209 the situation showed no signs of resolution, and Innocent threatened to excommunicate John if he did not acquiesce to Langton's appointment. When this threat failed, Innocent excommunicated the King in November 1209. Although theoretically a significant blow to John's legitimacy, this did not appear to worry the King greatly. Two of John's close allies, Emperor Otto IV and Count Raymond VI of Toulouse, had already suffered the same punishment themselves, and the significance of excommunication had been somewhat devalued. John simply tightened his existing measures and accrued significant sums from the income of vacant sees and abbeys: one 1213 estimate, for example, suggested the church had lost an estimated 100,000 marks (equivalent to £66,666 at the time) to John. Official figures suggest that around 14% of annual income from the English church was being appropriated by John each year. Innocent gave some dispensations as the crisis progressed. Monastic communities were allowed to celebrate Mass in private from 1209 onwards, and late in 1212 the Holy Viaticum for the dying was authorised. The rules on burials and lay access to churches appear to have been steadily circumvented, at least unofficially. Although the interdict was a burden to much of the population, it did not result in rebellion against John. By 1213, though, John was increasingly worried about the threat of French invasion. Some contemporary chroniclers suggested that in January Philip II of France had been charged with deposing John on behalf of the papacy, although it appears that Innocent merely prepared secret letters in case Innocent needed to claim the credit if Philip did successfully invade England. Under mounting political pressure, John finally negotiated terms for a reconciliation, and the papal terms for submission were accepted in the presence of the papal legate Pandulf Verraccio in May 1213 at the Templar Church at Dover. As part of the deal, John offered to surrender the Kingdom of England to the papacy for a feudal service of 1,000 marks (equivalent to £666 at the time) annually: 700 marks (£466) for England and 300 marks (£200) for Ireland, as well as recompensing the Church for revenue lost during the crisis. The agreement was formalised in the , or Golden Bull. This resolution produced mixed responses. Although some chroniclers felt that John had been humiliated by the sequence of events, there was little public reaction. Innocent benefited from the resolution of his long-standing English problem, but John probably gained more, as Innocent became a firm supporter of John for the rest of his reign, backing him in both domestic and continental policy issues. Innocent immediately turned against Philip, calling upon him to reject plans to invade England and to sue for peace. John paid some of the compensation money he had promised the Church, but he ceased making payments in late 1214, leaving two-thirds of the sum unpaid; Innocent appears to have conveniently forgotten this debt for the good of the wider relationship. Failure in France and the First Barons' War (1215–1216) Tensions and discontent Tensions between John and the barons had been growing for several years, as demonstrated by the 1212 plot against the King. Many of the disaffected barons came from the north of England; that faction was often labelled by contemporaries and historians as "the Northerners". The northern barons rarely had any personal stake in the conflict in France, and many of them owed large sums of money to John; the revolt has been characterised as "a rebellion of the king's debtors". Many of John's military household joined the rebels, particularly amongst those that John had appointed to administrative roles across England; their local links and loyalties outweighed their personal loyalty to John. Tension also grew across North Wales, where opposition to the 1211 treaty between John and Llywelyn was turning into open conflict. For some the appointment of Peter des Roches as justiciar was an important factor, as he was considered an "abrasive foreigner" by many of the barons. The failure of John's French military campaign in 1214 was probably the final straw that precipitated the baronial uprising during John's final years as king; James Holt describes the path to civil war as "direct, short and unavoidable" following the defeat at Bouvines. Failure of the 1214 French campaign In 1214 John began his final campaign to reclaim Normandy from Philip. He was optimistic, as he had successfully built up alliances with the Emperor Otto, Renaud of Boulogne and Ferdinand of Flanders; he was enjoying papal favour; and he had successfully built up substantial funds to pay for the deployment of his experienced army. Nonetheless, when John left for Poitou in February 1214, many barons refused to provide military service; mercenary knights had to fill the gaps. John's plan was to split Philip's forces by pushing north-east from Poitou towards Paris, whilst Otto, Renaud and Ferdinand, supported by William Longespée, marched south-west from Flanders. The first part of the campaign went well, with John outmanoeuvring the forces under the command of Prince Louis and retaking the county of Anjou by the end of June. John besieged the castle of Roche-au-Moine, a key stronghold, forcing Louis to give battle against John's larger army. The local Angevin nobles refused to advance with John; left at something of a disadvantage, John retreated back to La Rochelle. Shortly afterwards, King Philip won the hard-fought battle of Bouvines in the north against Otto and John's other allies, bringing an end to John's hopes of retaking Normandy. A peace agreement was signed in which John returned Anjou to Philip and paid him compensation; the truce was intended to last for six years. John arrived back in England in October. Pre-war tensions and Within a few months of John's return, rebel barons in the north and east of England were organising resistance to his rule. John held a council in London in January 1215 to discuss potential reforms and sponsored discussions in Oxford between his agents and the rebels during the spring. He appears to have been playing for time until Pope Innocent III could send letters giving him explicit papal support. This was particularly important for John, as a way of pressuring the barons but also as a way of controlling Stephen Langton, the Archbishop of Canterbury. In the meantime, John began to recruit fresh mercenary forces from Poitou, although some were later sent back to avoid giving the impression that John was escalating the conflict. The King announced his intent to become a crusader, a move which gave him additional political protection under church law. Letters of support from the Pope arrived in April but by then the rebel barons had organised. They congregated at Northampton in May and renounced their feudal ties to John, appointing Robert fitz Walter as their military leader. This self-proclaimed "Army of God" marched on London, taking the capital as well as Lincoln and Exeter. John's efforts to appear moderate and conciliatory had been largely successful, but once the rebels held London they attracted a fresh wave of defectors from John's royalist faction. John instructed Langton to organise peace talks with the rebel barons. John met the rebel leaders at Runnymede, near Windsor Castle, on 15 June 1215. Langton's efforts at mediation created a charter capturing the proposed peace agreement; it was later renamed , or "Great Charter". The charter went beyond simply addressing specific baronial complaints, and formed a wider proposal for political reform, albeit one focusing on the rights of free men, not serfs and unfree labour. It promised the protection of church rights, protection from illegal imprisonment, access to swift justice, new taxation only with baronial consent and limitations on scutage and other feudal payments. A council of twenty-five barons would be created to monitor and ensure John's future adherence to the charter, whilst the rebel army would stand down and London would be surrendered to the King. Neither John nor the rebel barons seriously attempted to implement the peace accord. The rebel barons suspected that the proposed baronial council would be unacceptable to John and that he would challenge the legality of the charter; they packed the baronial council with their own hardliners and refused to demobilise their forces or surrender London as agreed. Despite his promises to the contrary, John appealed to Innocent for help, observing that the charter compromised the Pope's rights under the 1213 agreement that had appointed him John's feudal lord. Innocent obliged; he declared the charter "not only shameful and demeaning, but illegal and unjust" and excommunicated the rebel barons. The failure of the agreement led rapidly to the First Barons' War. War with the barons The rebels made the first move in the war, seizing the strategic Rochester Castle, owned by Langton but left almost unguarded by the archbishop. John was well prepared for a conflict. He had stockpiled money to pay for mercenaries and ensured the support of the powerful marcher lords with their own feudal forces, such as William Marshal and Ranulf de Blondeville, 6th Earl of Chester. The rebels lacked the engineering expertise or heavy equipment necessary to assault the network of royal castles that cut off the northern rebel barons from those in the south. John's strategy was to isolate the rebel barons in London, protect his own supply lines to his key source of mercenaries in Flanders, prevent the French from landing in the south-east, and then win the war through slow attrition. John put off dealing with the badly deteriorating situation in North Wales, where Llywelyn the Great was leading a rebellion against the 1211 settlement. John's campaign started well. In November John retook Rochester Castle from rebel baron William d'Aubigny in a sophisticated assault. One chronicler had not seen "a siege so hard pressed or so strongly resisted", whilst historian Reginald Brown describes it as "one of the greatest [siege] operations in England up to that time". Having regained the south-east John split his forces, sending William Longespée to retake the north side of London and East Anglia, whilst John himself headed north via Nottingham to attack the estates of the northern barons. Both operations were successful and the majority of the remaining rebels were pinned down in London. In January 1216 John marched against Alexander II of Scotland, who had allied himself with the rebel cause. John took back Alexander's possessions in northern England in a rapid campaign and pushed up towards Edinburgh over a ten-day period. The rebel barons responded by inviting the French prince Louis to lead them: Louis had a claim to the English throne by virtue of his marriage to Blanche of Castile, a granddaughter of Henry II. Philip may have provided him with private support but refused to openly support Louis, who was excommunicated by Innocent for taking part in the war against John. Louis' planned arrival in England presented a significant problem for John, as the prince would bring with him naval vessels and siege engines essential to the rebel cause. Once John contained Alexander in Scotland, he marched south to deal with the challenge of the coming invasion. Prince Louis intended to land in the south of England in May 1216, and John assembled a naval force to intercept him. Unfortunately for John, his fleet was dispersed by bad storms and Louis landed unopposed in Kent. John hesitated and decided not to attack Louis immediately, either due to the risks of open battle or over concerns about the loyalty of his own men. Louis and the rebel barons advanced west and John retreated, spending the summer reorganising his defences across the rest of the kingdom. John saw several of his military household desert to the rebels, including his half-brother, William Longespée. By the end of the summer the rebels had regained the south-east of England and parts of the north. Death In September 1216, John began a fresh, vigorous attack. He marched from the Cotswolds, feigned an offensive to relieve the besieged Windsor Castle, and attacked eastwards around London to Cambridge to separate the rebel-held areas of Lincolnshire and East Anglia. From there he travelled north to relieve the rebel siege at Lincoln and back east to Lynn, probably to order further supplies from the continent. In Lynn, John contracted dysentery, which would ultimately prove fatal. Meanwhile, Alexander II invaded northern England again, taking Carlisle in August and then marching south to give homage to Prince Louis for his English possessions; John narrowly missed intercepting Alexander along the way. Tensions between Louis and the English barons began to increase, prompting a wave of desertions, including William Marshal's son William and William Longespée, who both returned to John's faction. Crown Jewels John returned west but is said to have lost much of his baggage train along the way. Roger of Wendover provides the most graphic account of this, suggesting that the King's belongings, including the English Crown Jewels, were lost as he crossed one of the tidal estuaries which empties into the Wash, being sucked in by quicksand and whirlpools. Accounts of the incident vary considerably between the various chroniclers and the exact location of the incident has never been confirmed; the losses may have involved only a few of his pack-horses. Modern historians assert that by October 1216 John faced a "stalemate", "a military situation uncompromised by defeat". John's illness grew worse and by the time he reached Newark Castle, Nottinghamshire, he was unable to travel any farther; he died on the night of 18/19 October. Numerous—probably fictitious—accounts circulated soon after his death that he had been killed by poisoned ale, poisoned plums or a "surfeit of peaches". His body was escorted south by a company of mercenaries and he was buried in Worcester Cathedral in front of the altar of St Wulfstan. A new sarcophagus with an effigy was made for him in 1232, in which his remains now rest. In his will, John ordered that his niece Eleanor, who might have had a claim to the throne of his successor, Henry III, never be released from prison. Legacy In the aftermath of John's death, William Marshal was declared the protector of the nine-year-old Henry III. The civil war continued until royalist victories at the battles of Lincoln and Dover in 1217. Louis gave up his claim to the English throne and signed the Treaty of Lambeth. The failed agreement was resuscitated by Marshal's administration and reissued in an edited form in 1217 as a basis for future government. Henry III continued his attempts to reclaim Normandy and Anjou until 1259, but John's continental losses and the consequent growth of Capetian power in the 13th century proved to mark a "turning point in European history". John's first wife, Isabella, Countess of Gloucester, was released from imprisonment in 1214; she remarried twice, and died in 1217. John's second wife, Isabella of Angoulême, left England for Angoulême soon after the king's death; she became a powerful regional leader, but largely abandoned the children she had had by John. Their eldest son, Henry III, ruled as King of England for the majority of the 13th century. Their other son, Richard of Cornwall, became a noted European leader and ultimately the King of the Romans in the Holy Roman Empire. Their daughter Joan became Queen of Scotland on her marriage to Alexander II. Another daughter, Isabella, was Holy Roman Empress as the wife of Emperor Frederick II. The youngest daughter, Eleanor, married William Marshal's son, also called William, and later the famous English rebel Simon de Montfort. By various mistresses John had eight, possibly nine, sons—Richard, Oliver, John, Geoffrey, Henry, Osbert Gifford, Eudes, Bartholomew and probably Philip—and two or three daughters—Joan, Maud and probably Isabel. Of these, Joan became the most famous, marrying Prince Llywelyn the Great of Wales. Historiography Historical interpretations of John have been subject to considerable change over the centuries. Medieval chroniclers provided the first contemporary, or near contemporary, histories of John's reign. One group of chroniclers wrote early in John's life, or around the time of his accession, including Richard of Devizes, William of Newburgh, Roger of Hoveden and Ralph de Diceto. These historians were generally unsympathetic to John's behaviour under Richard's rule, but slightly more positive towards the very earliest years of John's reign. Reliable accounts of the middle and later parts of John's reign are more limited, with Gervase of Canterbury and Ralph of Coggeshall writing the main accounts; neither of them were positive about John's performance as king. Much of John's later, negative reputation was established by two chroniclers writing after his death, Roger of Wendover and Matthew Paris, the latter claiming that John attempted conversion to Islam in exchange for military aid from the Almohad ruler Muhammad al-Nasir—a story modern historians consider untrue. In the 16th century political and religious changes altered the attitude of historians towards John. Tudor historians were generally favourably inclined towards the King, focusing on his opposition to the Papacy and his promotion of the special rights and prerogatives of a king. Revisionist histories written by John Foxe, William Tyndale and Robert Barnes portrayed John as an early Protestant hero, and Foxe included the King in his Book of Martyrs. John Speed's Historie of Great Britaine in 1632 praised John's "great renown" as a king; he blamed the bias of medieval chroniclers for the King's poor reputation. By the Victorian period in the 19th century, historians were more inclined to draw on the judgements of the chroniclers and to focus on John's moral personality. Kate Norgate, for example, argued that John's downfall had been due not to his failure in war or strategy, but due to his "almost superhuman wickedness", whilst James Ramsay blamed John's family background and his cruel personality for his downfall. Historians in the "Whiggish" tradition, focusing on documents such as the Domesday Book and , trace a progressive and universalist course of political and economic development in England over the medieval period. These historians were often inclined to see John's reign, and his signing of in particular, as a positive step in the constitutional development of England, despite the flaws of the King himself. Winston Churchill, for example, argued that "[w]hen the long tally is added, it will be seen that the British nation and the English-speaking world owe far more to the vices of John than to the labours of virtuous sovereigns". In the 1940s, new interpretations of John's reign began to emerge, based on research into the record evidence of his reign, such as pipe rolls, charters, court documents and similar primary records. Notably, an essay by Vivian Galbraith in 1945 proposed a "new approach" to understanding the ruler. The use of recorded evidence was combined with an increased scepticism about two of the most colourful chroniclers of John's reign, Roger of Wendover and Matthew Paris. In many cases the detail provided by these chroniclers, both writing after John's death, was challenged by modern historians. Interpretations of the and the role of the rebel barons in 1215 have been significantly revised: although the charter's symbolic, constitutional value for later generations is unquestionable, in the context of John's reign most historians now consider it a failed peace agreement between "partisan" factions. There has been increasing debate about the nature of John's Irish policies. Specialists in Irish medieval history, such as Sean Duffy, have challenged the conventional narrative established by Lewis Warren, suggesting that Ireland was less stable by 1216 than was previously supposed. Most historians today, including John's recent biographers Ralph Turner and Lewis Warren, argue that John was an unsuccessful monarch, but note that his failings were exaggerated by 12th- and 13th-century chroniclers. Jim Bradbury notes the current consensus that John was a "hard-working administrator, an able man, an able general", albeit, as Turner suggests, with "distasteful, even dangerous personality traits", including pettiness, spitefulness and cruelty. John Gillingham, author of a major biography of Richard I, follows this line too, although he considers John a less effective general than do Turner or Warren, and describes him "one of the worst kings ever to rule England". Bradbury takes a moderate line, but suggests that in recent years modern historians have been overly lenient towards John's numerous faults. Popular historian Frank McLynn maintains a counter-revisionist perspective on John, arguing that the King's modern reputation amongst historians is "bizarre", and that as a monarch John "fails almost all those [tests] that can be legitimately set". According to C. Warren Hollister, "The dramatic ambivalence of his personality, the passions that he stirred among his own contemporaries, the very magnitude of his failures, have made him an object of endless fascination to historians and biographers." Popular representations Popular representations of John first began to emerge during the Tudor period, mirroring the revisionist histories of the time. The anonymous play The Troublesome Reign of King John portrayed the King as a "proto-Protestant martyr", similar to that shown in John Bale's morality play Kynge Johan, in which John attempts to save England from the "evil agents of the Roman Church". By contrast, Shakespeare's King John, a relatively anti-Catholic play that draws on The Troublesome Reign for its source material, offers a more "balanced, dual view of a complex monarch as both a proto-Protestant victim of Rome's machinations and as a weak, selfishly motivated ruler". Anthony Munday's play The Downfall and The Death of Robert Earl of Huntington portrays many of John's negative traits, but adopts a positive interpretation of the King's stand against the Roman Catholic Church, in line with the contemporary views of the Tudor monarchs. By the middle of the 17th century, plays such as Robert Davenport's King John and Matilda, although based largely on the earlier Elizabethan works, were transferring the role of Protestant champion to the barons and focusing more on the tyrannical aspects of John's behaviour. Nineteenth-century fictional depictions of John were heavily influenced by Sir Walter Scott's historical romance, Ivanhoe, which presented "an almost totally unfavourable picture" of the King; the work drew on 19th-century histories of the period and on Shakespeare's play. Scott's work influenced the late-19th-century children's writer Howard Pyle's book The Merry Adventures of Robin Hood, which in turn established John as the principal villain within the traditional Robin Hood narrative. During the 20th century, John was normally depicted in fictional books and films alongside Robin Hood. Sam De Grasse's role as John in the black-and-white 1922 film version shows John committing numerous atrocities and acts of torture. Claude Rains played John in the 1938 colour version alongside Errol Flynn, starting a trend for films to depict John as an "effeminate ... arrogant and cowardly stay-at-home". The character of John acts either to highlight the virtues of King Richard, or contrasts with the Sheriff of Nottingham, who is usually the "swashbuckling villain" opposing Robin. An extreme version of this trend can be seen in the 1973 Disney cartoon version, for example, which depicts John, voiced by Peter Ustinov, as a "cowardly, thumbsucking lion". Popular works that depict John beyond the Robin Hood legends, such as James Goldman's play and later film, The Lion in Winter, set in 1183, commonly present him as an "effete weakling", in this instance contrasted with the more masculine Henry II, or as a tyrant, as in A. A. Milne's poem for children, "King John's Christmas". Issue John and Isabella of Angoulême had five children: Henry III, King of England (1 October 1207 – 16 November 1272) Richard, King of the Romans (5 January 1209 – 2 April 1272) Joan, Queen of Scotland (22 July 1210 – 4 March 1238) Isabella, Holy Roman Empress (1214 – 1 December 1241) Eleanor, Countess of Pembroke (1215 – 13 April 1275) John had several mistresses, including one named Suzanne. His known illegitimate children are: Richard FitzRoy ( – June 1246), whose mother was Adela, John's first cousin Joan, Lady of Wales ( – February 1237), also known by her Welsh name of Siwan John (fl. 1201), who became a clerk Geoffrey (died 1205), held the honour of Perche Oliver fitz Regis (bef. 1199 – 1218/1219), whose mother was Hawise, sister of Fulk FitzWarin Osbert Giffard Genealogical table Notes References Bibliography in in in in in in in in in in in in in in in in in in External links John Lackland at the official website of the British monarchy |- |- |- |- 1166 births 1216 deaths 12th-century English monarchs 13th-century English monarchs 12th-century Dukes of Normandy 13th-century Dukes of Normandy Deaths from dysentery Earls of Cornwall English people of French descent English people of Scottish descent High Sheriffs of Somerset House of Plantagenet People temporarily excommunicated by the Catholic Church People from Oxford Robin Hood characters People of the Barons' Wars Burials at Worcester Cathedral House of Anjou 13th-century peers of France Earls of Gloucester Infectious disease deaths in England Children of Henry II of England Norman warriors Anglo-Normans Lords of Glamorgan Proto-Protestants
16698
https://en.wikipedia.org/wiki/Politics%20of%20Kyrgyzstan
Politics of Kyrgyzstan
The politics of Kyrgyzstan, officially known as the Kyrgyz Republic, takes place in the framework of a presidential system representative democratic republic, whereby the President is head of state and the Chairman of the Cabinet of Ministers is head of government. Executive power is exercised by the government. Legislative power is vested in both the government and parliament. Unlike its authoritarian neighbors – Kazakhstan, Tajikistan, Turkmenistan, and Uzbekistan – Kyrgyzstan has had a more pluralistic political system since the collapse of the Soviet Union. Kyrgyzstan has swung between authoritarian and democratic forms of government. Three authoritarian presidents have been ousted from office since 2005 due to popular protests. Political history In the first years of Kyrgyzstan's full independence, President Askar Akayev appeared wholeheartedly committed to the reform process. However, despite the backing of major Western donors, including the International Monetary Fund (IMF), Kyrgyzstan had consequential economic difficulties from the outset. These came mainly as a result of the breakup of the Soviet trade bloc, which impeded the Republic's smooth transfer to a free-market economy. In 1993, allegations of corruption against Akayev's closest political associates blossomed into a major scandal. One of those accused of improprieties was Vice President Feliks Kulov, who resigned for ethical reasons in December. Following Kulov's resignation, Akayev dismissed the government and called upon the last communist premier, Apas Djumagulov, to form a new one. In January 1994, Akayev initiated a referendum asking for a renewed mandate to complete his term of office. He received 96.2% of the vote. A new Constitution was passed by the Parliament in May 1993. In 1994, however, the Parliament failed to produce a quorum for its last scheduled session prior to the expiration of its term (February 1995). President Akayev was widely accused of having manipulated a boycott by a majority of the parliamentarians. Akayev, in turn, asserted that the communists had caused a political crisis by preventing the legislature from fulfilling its role. Akayev scheduled an October 1994 referendum, overwhelmingly approved by voters, that proposed two amendments to the Constitution, one that would allow the Constitution to be amended by means of a referendum, and the other creating a new bicameral parliament called the Jogorku Keņesh. Elections for the two legislative chambers – a 35-seat full-time assembly and a 70-seat part-time assembly – were held in February 1995 after campaigns considered remarkably free and open by most international observers, although the election-day proceedings were marred by widespread irregularities. Independent candidates won most of the seats, suggesting that personalities prevailed over ideologies. The new Parliament convened its initial session in March 1995. One of its first orders of business was the approval of the precise constitutional language on the role of the legislature. Kyrgyzstan's independent political parties competed in the 1996 parliamentary elections. A February 1996 referendum – in violation of the Constitution and the law on referendums – amended the Constitution to give President Akayev more power. It also removed the clause that parliamentarians be directly elected by universal suffrage. Although the changes gave the President the power to dissolve Parliament, it also more clearly defined Parliament's powers. Since that time, Parliament has demonstrated real independence from the executive branch. An October 1998 referendum approved constitutional changes, including increasing the number of deputies in the upper house, reducing the number of deputies in the lower house, rolling back Parliamentary immunity, reforming land tender rules, and reforming the state budget. Two rounds of Parliamentary elections were held on 20 February 2000 and 12 March 2000. With the full backing of the United States, the Organization for Security and Co-operation in Europe (OSCE) reported that the elections failed to comply with commitments to free and fair elections and hence were invalid. Questionable judicial proceedings against opposition candidates and parties limited the choice of candidates available to Kyrgyz voters, while state-controlled media reported favorably on official candidates only and government officials put pressure on independent media outlets that favored the opposition. In 2002 Azimbek Beknazarov, a leading opposition figure, was imprisoned by the local authorities, in what many believe to be politically motivated circumstances. This led to protests resulting in clashes with police forces, culminating in the death of five people in Jalal-Abad. As May approached the authorities further extended their hold on power, imprisoning the vocal former presidential ally, Feliks Kulov, to ten years for alleged "abuses of office". During the same month the entire government resigned, accepting blame for the loss of life during the protests earlier in the year. A new government led by Nikolay Tanayev was then formed. In November the President faced yet more protests, as the opposition announced it would march on the capital and demand his resignation. The police reacted by arresting large amounts of demonstrators, further adding to international disapproval at the authoritarian nature of Akayev's government. By June 2003, the lower house of Parliament announced that President Akayev and two other leaders of Kyrgyzstan, from the Soviet era, would be given lifetime immunity from prosecution, raising the prospect of Akayev finally stepping down. In 2005, following disputed results of the 2005 parliamentary elections, Kyrgyzstan was thrown into a state of political turmoil, with different parties claiming that they were the legitimate government. On 10 July 2005 interim President and opposition People's Movement leader Kurmanbek Bakiyev won the presidential election in a landslide victory. (See: Tulip Revolution). In 2006, Bakiyev faced a political crisis as thousands of people demonstrated in a series of protests in Bishkek. They accused him of reneging on promised constitutional reforms limiting presidential power and giving more authority to the parliament and cabinet. They also accused him of failing to eradicate corruption, crime, and poverty. Bakiyev in turn accused the opposition of plotting a coup against him. Several parliamentarians had been killed during the political unrest. Presidential elections, originally expected in 2010, were rescheduled for 23 July 2009. President Bakiyev was widely expected to retain his mandate, while the opposition United People's Movement (OND) announced on 20 April 2009 that it would field a single candidate – Social Democratic Party leader Almaz Atambayev. The election turnout was reported at 79.3%. As of 00:45 local time in Kyrgyzstan on 25 July 2009 (with 2058 of 2330 polling districts reporting), Bakiyev had won the election with 83.8% of the vote. In assessing the election, the OSCE stated that Bakiyev had gained an "unfair advantage" and that media bias "did not allow voters to make an informed choice." Additionally, they found that the election was "marred by many problems and irregularities", citing ballot stuffing and problems with the counting of votes. On polling day Atambayev withdrew his candidacy claiming widespread fraud, stating "due to massive, unprecedented violations, we consider these elections illegitimate and a new election should be held." An opposition rally of 1,000 people in Balykchy on election day was broken up by riot police. The arrest of an opposition figure on 6 April 2010 in the town of Talas led opposition supporters to protest. The protestors took control of a governmental building, demanding a new government. Riot police were sent from Bishkek, and managed to temporarily regain control of the building. Later the same day several more opposition figures were arrested, while the government claimed to have regained control of the situation. The following day, however, hundreds of opposition supporters gathered in Bishkek and marched on the government headquarters. Security personnel attempted to disperse the protestors with the use of stun grenades and live rounds, at the cost of dozens of lives. The protests continued, however, resulting in the flight of President Bakiyev to his southern stronghold of Jalalabad, and the freeing later the same day of the arrested opposition figures. A new government was formed under opposition leader Roza Otunbayeva, while Bakiyev remained for several days in southern Kyrgyzstan, before fleeing to Belarus, where he was given asylum by President Lukashenko. The new interim government held consultations on a new constitution, intended to increase the powers of the parliament and reduce those of the president. A referendum was held on the resulting document on 27 June 2010, and was approved by over 90% of voters, with a turnout of 72%. Elections were subsequently held on 10 October 2010. These elections resulted in five parties reaching the 5% threshold necessary to enter parliament. Presidential elections were held in 2011, resulting in the victory of Almazbek Atambayev. In 2017, he endorsed Prime minister Sooronbay Jeenbekov for President, who won that year's presidential contest. After Atambayev retired from presidency, he began criticizing Jeenbekov and their relations worsened over time. Soon Atambayev was accused of corruption by the Jeenbekov administration. Clashes grew between security forces and Atambayev supporters, exacerbating political turmoil. Political insurgency in Kyrgyzstan amplified following controversy about the conduct of October 2020 Kyrgyz parliamentary election, where only four parties met the 7% threshold to achieve parliamentary representation, three of whom were closely aligned with the government. Opposition supporters claimed that these elections were tainted by vote buying and many other forms of irregularities. The protests resulted in en masse resignation of government officials. In October 2020, President Sooronbay Jeenbekov resigned after protests caused by irregularities in parliamentary elections on 4 October 2020. In January 2021, Sadyr Japarov was elected as the new president after winning the presidential election by landslide. In April 2021, the majority of voters approved in the constitutional referendum a new constitution that would give new powers to the president, strengthening significantly the presidency. Executive branch |President |Sadyr Japarov |Mekenchil |28 January 2021 |- |Chairman of the Cabinet of Ministers |Akylbek Japarov |Ar-Namys |12 October 2021 |} The president is elected by popular vote for a maximum of two five-year terms. The chairman of the Cabinet of Ministers is appointed by the president and approved by the Supreme Council. Cabinet of Ministers Akylbek Japarov Government Since 13 October 2021: Chairman of the cabinet of ministers – Akylbek Japarov First Deputy Chairman – Arzybek Kojoshev Deputy Chairman – Edil Baisalov Deputy Chairman – Bakyt Torobaev Deputy Chairman, Head of SCNS – Kamchybek Tashiev Minister of Foreign Affairs – Jeenbek Kulubaev Minister of Agriculture – Askar Janybekov Minister of Digital Development – Talantbek Imanov Minister of the Economy – Daniyar Amangeldiev Minister of Education and Science – Almazbek Beishenaliev Minister of Emergency Situations – Boobek Ajikeev Minister of Culture, Information and Tourism – Azamat Jamankulov Minister of Finance – Almaz Baketaev Minister of Health Care – Jarkynbek Kasymbekov Minister of Internal Affairs – Ulan Niyazbekov Minister of Justice – Ayaz Baetov Minister of Transport and Communication – Erkinbek Osoev Minister of Energy – Doskul Bekmurzaev Minister of Labor, Welfare and Migration – Kudaibergen Bazarbaev Minister of Defence – Baktybek Bekbolotov Minister of Natural Resources, Environment and Technical Supervision – Dinara Kutmanova Legislative branch In the Soviet era, Kyrgyzstan had a unicameral legislature which was replaced in 1995 by the bicameral Supreme Council (Joghorku Keneš). The Supreme Council consisted of the Assembly of People's Representatives (45 seats; members were elected by popular vote from single member constituencies) and the Legislative Assembly (60 seats; 45 members of which were elected by popular vote from single member constituencies, and 15 of which were from national party lists on a proportional basis with a 5% threshold). All legislative terms were five years. In 2005, as part of the 2005 election process and in accordance with a 2003 referendum, the Parliament again became unicameral. The Legislative Assembly (Myizam Chygaruu Jyiyny) had 75 members, elected for five-year terms from single-seat constituencies. However, because of the political unrest, a new constitutional referendum was held on 21 October 2007 which approved a new electoral system, enlarged the parliament to 90 members and introducing party-list voting. Party-list voting is a proportional representation system of voting, where candidates are selected from central party lists rather than locally elected. Early parliamentary elections were held on 16 December 2007. Political parties and elections 2017 presidential elections 2015 parliamentary elections Political pressure groups and leaders Council of Free Trade Unions Kyrgyz Committee on Human Rights – Ramazan Dyryldayev National Unity Democratic Movement Union of Entrepreneurs Central Asian Free Market Institute Judicial branch Although the constitution provides for an independent judiciary, Kyrgyzstan's court system is widely seen as under the influence of the prosecutor's office. Low salaries make the bribery of judges commonplace. Most cases originate in local courts; they then can move via the appeals process to municipal or regional courts, with the Supreme Court the final court of appeals. Property and family law disputes and low-level criminal cases are heard by traditional elders' courts, which are loosely supervised by the prosecutor's office. Economic disputes and military cases are heard in specialized courts. The constitutional amendments of 2003 expanded the scope of the Supreme Court in civil, criminal, and administrative proceedings. Many protections of Western jurisprudence are not present in Kyrgyzstan's system, which retains many features of the Soviet system. The right to counsel and the presumption of innocence of the accused are guaranteed by law but often not practiced. There is no trial by jury. Reform legislation under consideration in 2006 would establish a jury system and bolster the independence of the judicial branch. The Prosecutor General's Office of Kyrgyzstan oversees the enforcement of the Kyrgyz legal system and the activities of law enforcement agencies and the sentencing of criminals in concert with the judiciary. Administrative divisions Kyrgyzstan is divided into seven Regions (oblustar, singular – oblus) and two region-level cities* (shaar): Batken Region (Batken) Bishkek* Chüy Region (Bishkek) Issyk-Kul Region (Karakol) Jalal-Abad Region (Jalal-Abad) Naryn Region (Naryn) Osh* Osh Region (Osh) Talas Region (Talas) note: administrative center names in parentheses International organization participation AsDB CIS EAPC EBRD ECO FAO IBRD ICAO ICCt (signatory) ICRM IDA IDB IFAD IFC IFRCS ILO IMF INOGATE Interpol IOC IOM ISO (correspondent) ITU NAM (observer) OIC OPCW OSCE PCA PFP SCO UN UNAMSIL UNCTAD (UNECE) United Nations Economic Commission for Europe UNESCO UNIDO UNMIK UPU WCO WFTU WHO WIPO WMO WToO WTO UNICEF Notes External links Erik Herron's Guide to Politics in East Central Europe and Eurasia Kyrgyz Report offers ongoing analysis, including an analysis of the clusters of political forces in the country and the roots of the country’s political problems Government of Kyrgyz Republic bn:কিরগিজিস্তান#রাজনীতি
16848
https://en.wikipedia.org/wiki/Kurtosis
Kurtosis
In probability theory and statistics, kurtosis (from , kyrtos or kurtos, meaning "curved, arching") is a measure of the "tailedness" of the probability distribution of a real-valued random variable. Like skewness, kurtosis describes a particular aspect of a probability distribution. There are different ways to quantify kurtosis for a theoretical distribution, and there are corresponding ways of estimating it using a sample from a population. Different measures of kurtosis may have different interpretations. The standard measure of a distribution's kurtosis, originating with Karl Pearson, is a scaled version of the fourth moment of the distribution. This number is related to the tails of the distribution, not its peak; hence, the sometimes-seen characterization of kurtosis as "peakedness" is incorrect. For this measure, higher kurtosis corresponds to greater extremity of deviations (or outliers), and not the configuration of data near the mean. It is common to compare the excess kurtosis (defined below) of a distribution to 0, which is the excess kurtosis of any univariate normal distribution. Distributions with negative excess kurtosis are said to be platykurtic, although this does not imply the distribution is "flat-topped" as is sometimes stated. Rather, it means the distribution produces fewer and/or less extreme outliers than the normal distribution. An example of a platykurtic distribution is the uniform distribution, which does not produce outliers. Distributions with a positive excess kurtosis are said to be leptokurtic. An example of a leptokurtic distribution is the Laplace distribution, which has tails that asymptotically approach zero more slowly than a Gaussian, and therefore produces more outliers than the normal distribution. It is common practice to use excess kurtosis, which is defined as Pearson's kurtosis minus 3, to provide a simple comparison to the normal distribution. Some authors and software packages use "kurtosis" by itself to refer to the excess kurtosis. For clarity and generality, however, this article explicitly indicates where non-excess kurtosis is meant. Alternative measures of kurtosis are: the L-kurtosis, which is a scaled version of the fourth L-moment; measures based on four population or sample quantiles. These are analogous to the alternative measures of skewness that are not based on ordinary moments. Pearson moments The kurtosis is the fourth standardized moment, defined as where μ4 is the fourth central moment and σ is the standard deviation. Several letters are used in the literature to denote the kurtosis. A very common choice is κ, which is fine as long as it is clear that it does not refer to a cumulant. Other choices include γ2, to be similar to the notation for skewness, although sometimes this is instead reserved for the excess kurtosis. The kurtosis is bounded below by the squared skewness plus 1: where μ3 is the third central moment. The lower bound is realized by the Bernoulli distribution. There is no upper limit to the kurtosis of a general probability distribution, and it may be infinite. A reason why some authors favor the excess kurtosis is that cumulants are extensive. Formulas related to the extensive property are more naturally expressed in terms of the excess kurtosis. For example, let X1, ..., Xn be independent random variables for which the fourth moment exists, and let Y be the random variable defined by the sum of the Xi. The excess kurtosis of Y is where is the standard deviation of . In particular if all of the Xi have the same variance, then this simplifies to The reason not to subtract 3 is that the bare fourth moment better generalizes to multivariate distributions, especially when independence is not assumed. The cokurtosis between pairs of variables is an order four tensor. For a bivariate normal distribution, the cokurtosis tensor has off-diagonal terms that are neither 0 nor 3 in general, so attempting to "correct" for an excess becomes confusing. It is true, however, that the joint cumulants of degree greater than two for any multivariate normal distribution are zero. For two random variables, X and Y, not necessarily independent, the kurtosis of the sum, X + Y, is Note that the fourth-power binomial coefficients (1, 4, 6, 4, 1) appear in the above equation. Interpretation The exact interpretation of the Pearson measure of kurtosis (or excess kurtosis) used to be disputed, but is now settled. As Westfall notes in 2014, "...its only unambiguous interpretation is in terms of tail extremity; i.e., either existing outliers (for the sample kurtosis) or propensity to produce outliers (for the kurtosis of a probability distribution)." The logic is simple: Kurtosis is the average (or expected value) of the standardized data raised to the fourth power. Standardized values that are less than 1 (i.e., data within one standard deviation of the mean, where the "peak" would be) contribute virtually nothing to kurtosis, since raising a number that is less than 1 to the fourth power makes it closer to zero. The only data values (observed or observable) that contribute to kurtosis in any meaningful way are those outside the region of the peak; i.e., the outliers. Therefore, kurtosis measures outliers only; it measures nothing about the "peak". Many incorrect interpretations of kurtosis that involve notions of peakedness have been given. One is that kurtosis measures both the "peakedness" of the distribution and the heaviness of its tail. Various other incorrect interpretations have been suggested, such as "lack of shoulders" (where the "shoulder" is defined vaguely as the area between the peak and the tail, or more specifically as the area about one standard deviation from the mean) or "bimodality". Balanda and MacGillivray assert that the standard definition of kurtosis "is a poor measure of the kurtosis, peakedness, or tail weight of a distribution" and instead propose to "define kurtosis vaguely as the location- and scale-free movement of probability mass from the shoulders of a distribution into its center and tails". Moors' interpretation In 1986 Moors gave an interpretation of kurtosis. Let where X is a random variable, μ is the mean and σ is the standard deviation. Now by definition of the kurtosis , and by the well-known identity . The kurtosis can now be seen as a measure of the dispersion of Z2 around its expectation. Alternatively it can be seen to be a measure of the dispersion of Z around +1 and −1. κ attains its minimal value in a symmetric two-point distribution. In terms of the original variable X, the kurtosis is a measure of the dispersion of X around the two values μ ± σ. High values of κ arise in two circumstances: where the probability mass is concentrated around the mean and the data-generating process produces occasional values far from the mean, where the probability mass is concentrated in the tails of the distribution. Excess kurtosis The excess kurtosis is defined as kurtosis minus 3. There are 3 distinct regimes as described below. Mesokurtic Distributions with zero excess kurtosis are called mesokurtic, or mesokurtotic. The most prominent example of a mesokurtic distribution is the normal distribution family, regardless of the values of its parameters. A few other well-known distributions can be mesokurtic, depending on parameter values: for example, the binomial distribution is mesokurtic for . Leptokurtic A distribution with positive excess kurtosis is called leptokurtic, or leptokurtotic. "Lepto-" means "slender". In terms of shape, a leptokurtic distribution has fatter tails. Examples of leptokurtic distributions include the Student's t-distribution, Rayleigh distribution, Laplace distribution, exponential distribution, Poisson distribution and the logistic distribution. Such distributions are sometimes termed super-Gaussian. Platykurtic A distribution with negative excess kurtosis is called platykurtic, or platykurtotic. "Platy-" means "broad". In terms of shape, a platykurtic distribution has thinner tails. Examples of platykurtic distributions include the continuous and discrete uniform distributions, and the raised cosine distribution. The most platykurtic distribution of all is the Bernoulli distribution with p = 1/2 (for example the number of times one obtains "heads" when flipping a coin once, a coin toss), for which the excess kurtosis is −2. Graphical examples The Pearson type VII family The effects of kurtosis are illustrated using a parametric family of distributions whose kurtosis can be adjusted while their lower-order moments and cumulants remain constant. Consider the Pearson type VII family, which is a special case of the Pearson type IV family restricted to symmetric densities. The probability density function is given by where a is a scale parameter and m is a shape parameter. All densities in this family are symmetric. The kth moment exists provided m > (k + 1)/2. For the kurtosis to exist, we require m > 5/2. Then the mean and skewness exist and are both identically zero. Setting a2 = 2m − 3 makes the variance equal to unity. Then the only free parameter is m, which controls the fourth moment (and cumulant) and hence the kurtosis. One can reparameterize with , where is the excess kurtosis as defined above. This yields a one-parameter leptokurtic family with zero mean, unit variance, zero skewness, and arbitrary non-negative excess kurtosis. The reparameterized density is In the limit as one obtains the density which is shown as the red curve in the images on the right. In the other direction as one obtains the standard normal density as the limiting distribution, shown as the black curve. In the images on the right, the blue curve represents the density with excess kurtosis of 2. The top image shows that leptokurtic densities in this family have a higher peak than the mesokurtic normal density, although this conclusion is only valid for this select family of distributions. The comparatively fatter tails of the leptokurtic densities are illustrated in the second image, which plots the natural logarithm of the Pearson type VII densities: the black curve is the logarithm of the standard normal density, which is a parabola. One can see that the normal density allocates little probability mass to the regions far from the mean ("has thin tails"), compared with the blue curve of the leptokurtic Pearson type VII density with excess kurtosis of 2. Between the blue curve and the black are other Pearson type VII densities with γ2 = 1, 1/2, 1/4, 1/8, and 1/16. The red curve again shows the upper limit of the Pearson type VII family, with (which, strictly speaking, means that the fourth moment does not exist). The red curve decreases the slowest as one moves outward from the origin ("has fat tails"). Other well-known distributions Several well-known, unimodal, and symmetric distributions from different parametric families are compared here. Each has a mean and skewness of zero. The parameters have been chosen to result in a variance equal to 1 in each case. The images on the right show curves for the following seven densities, on a linear scale and logarithmic scale: D: Laplace distribution, also known as the double exponential distribution, red curve (two straight lines in the log-scale plot), excess kurtosis = 3 S: hyperbolic secant distribution, orange curve, excess kurtosis = 2 L: logistic distribution, green curve, excess kurtosis = 1.2 N: normal distribution, black curve (inverted parabola in the log-scale plot), excess kurtosis = 0 C: raised cosine distribution, cyan curve, excess kurtosis = −0.593762... W: Wigner semicircle distribution, blue curve, excess kurtosis = −1 U: uniform distribution, magenta curve (shown for clarity as a rectangle in both images), excess kurtosis = −1.2. Note that in these cases the platykurtic densities have bounded support, whereas the densities with positive or zero excess kurtosis are supported on the whole real line. One cannot infer that high or low kurtosis distributions have the characteristics indicated by these examples. There exist platykurtic densities with infinite support, e.g., exponential power distributions with sufficiently large shape parameter b and there exist leptokurtic densities with finite support. e.g., a distribution that is uniform between −3 and −0.3, between −0.3 and 0.3, and between 0.3 and 3, with the same density in the (−3, −0.3) and (0.3, 3) intervals, but with 20 times more density in the (−0.3, 0.3) interval Also, there exist platykurtic densities with infinite peakedness, e.g., an equal mixture of the beta distribution with parameters 0.5 and 1 with its reflection about 0.0 and there exist leptokurtic densities that appear flat-topped, e.g., a mixture of distribution that is uniform between -1 and 1 with a T(4.0000001) Student's t-distribution, with mixing probabilities 0.999 and 0.001. Sample kurtosis Definitions A natural but biased estimator For a sample of n values, a method of moments estimator of the population excess kurtosis can be defined as where m4 is the fourth sample moment about the mean, m2 is the second sample moment about the mean (that is, the sample variance), xi is the ith value, and is the sample mean. This formula has the simpler representation, where the values are the standardized data values using the standard deviation defined using n rather than n − 1 in the denominator. For example, suppose the data values are 0, 3, 4, 1, 2, 3, 0, 2, 1, 3, 2, 0, 2, 2, 3, 2, 5, 2, 3, 999. Then the values are −0.239, −0.225, −0.221, −0.234, −0.230, −0.225, −0.239, −0.230, −0.234, −0.225, −0.230, −0.239, −0.230, −0.230, −0.225, −0.230, −0.216, −0.230, −0.225, 4.359 and the values are 0.003, 0.003, 0.002, 0.003, 0.003, 0.003, 0.003, 0.003, 0.003, 0.003, 0.003, 0.003, 0.003, 0.003, 0.003, 0.003, 0.002, 0.003, 0.003, 360.976. The average of these values is 18.05 and the excess kurtosis is thus 18.05 − 3 = 15.05. This example makes it clear that data near the "middle" or "peak" of the distribution do not contribute to the kurtosis statistic, hence kurtosis does not measure "peakedness". It is simply a measure of the outlier, 999 in this example. Standard unbiased estimator Given a sub-set of samples from a population, the sample excess kurtosis above is a biased estimator of the population excess kurtosis. An alternative estimator of the population excess kurtosis, which is unbiased in random samples of a normal distribution, is defined as follows: where k4 is the unique symmetric unbiased estimator of the fourth cumulant, k2 is the unbiased estimate of the second cumulant (identical to the unbiased estimate of the sample variance), m4 is the fourth sample moment about the mean, m2 is the second sample moment about the mean, xi is the ith value, and is the sample mean. This adjusted Fisher–Pearson standardized moment coefficient is the version found in Excel and several statistical packages including Minitab, SAS, and SPSS. Unfortunately, in nonnormal samples is itself generally biased. Upper bound An upper bound for the sample kurtosis of n (n > 2) real numbers is where is the corresponding sample skewness. Variance under normality The variance of the sample kurtosis of a sample of size n from the normal distribution is Stated differently, under the assumption that the underlying random variable is normally distributed, it can be shown that . Applications The sample kurtosis is a useful measure of whether there is a problem with outliers in a data set. Larger kurtosis indicates a more serious outlier problem, and may lead the researcher to choose alternative statistical methods. D'Agostino's K-squared test is a goodness-of-fit normality test based on a combination of the sample skewness and sample kurtosis, as is the Jarque–Bera test for normality. For non-normal samples, the variance of the sample variance depends on the kurtosis; for details, please see variance. Pearson's definition of kurtosis is used as an indicator of intermittency in turbulence. It is also used in magnetic resonance imaging to quantify non-Gaussian diffusion. A concrete example is the following lemma by He, Zhang, and Zhang: Assume a random variable has expectation , variance and kurtosis . Assume we sample many independent copies. Then . This shows that with many samples, we will see one that is above the expectation with probability at least . In other words: If the kurtosis is large, we might see a lot values either all below or above the mean. Kurtosis convergence Applying band-pass filters to digital images, kurtosis values tend to be uniform, independent of the range of the filter. This behavior, termed kurtosis convergence, can be used to detect image splicing in forensic analysis. Other measures A different measure of "kurtosis" is provided by using L-moments instead of the ordinary moments. See also Kurtosis risk Maximum entropy probability distribution References Further reading Alternative source (Comparison of kurtosis estimators) External links Kurtosis calculator Free Online Software (Calculator) computes various types of skewness and kurtosis statistics for any dataset (includes small and large sample tests).. Kurtosis on the Earliest known uses of some of the words of mathematics Celebrating 100 years of Kurtosis a history of the topic, with different measures of kurtosis. Moment (mathematics) Statistical deviation and dispersion
17001
https://en.wikipedia.org/wiki/Kohlrabi
Kohlrabi
Kohlrabi (pronounced ; scientific name Brassica oleracea Gongylodes Group), also called German turnip or turnip cabbage, is a biennial vegetable, a low, stout cultivar of wild cabbage. It is a cultivar of the same species as cabbage, broccoli, cauliflower, kale, Brussels sprouts, collard greens, Savoy cabbage, and gai lan. It can be eaten raw or cooked. Edible preparations are made with both the stem and the leaves. Despite its common names, it is not the same species as turnip, although both are in the genus Brassica. Etymology The name comes from the German Kohl ("cabbage") plus Rübe ~ Rabi (Swiss German variant) ("turnip"), because the swollen stem resembles the latter. Its Group name Gongylodes (or lowercase and italicized gongylodes or gongyloides as a variety name) means "roundish" in Greek, from (, ‘round’). In Iran, it is called kalam qomi or kalam qomri () and is used for various dishes, including a type of soup. In the northern part of Vietnam, it is called (from French chou-rave); in eastern parts of India (West Bengal) and Bangladesh, it is called ōl kapi. It is also found in the Kashmir Valley in Northern India and is there known as monj-hakh, monj being the round part, and hakh being the leafy part. In the hilly areas of Jammu region and its adjoining areas it is known as kādam () and is eaten throughout the year. It is called nol khol in Northern India, navalkōl () in Maharashtra, nūlkōl () in Tamil, nūl kōl () in Telugu, navilu kōsu () in Karnataka and in Sri Lanka as knol khol (turnip cabbage). In Cyprus it is known as kouloumpra (). It is eaten in the Czech Republic under name kedlubna, while in Slovakia it is known as kaleráb. In Romania, it is the gulie or cărălabă, which is similar to the Polish kalarepa and to the Hungarian karalábé, all of these last three denominations being adaptations of the German word Kohlrabi. History The first European written record is by the botanist Mattioli in 1554 who wrote that it had “come lately into Italy”. By the end of the 16th century, kohlrabi spread to North Europe and was being grown in Austria, Germany, England, Italy, Spain, Tripoli and parts of the eastern Mediterranean. Description Kohlrabi has been created by artificial selection for lateral meristem growth (a swollen, nearly spherical shape); its origin in nature is the same as that of cabbage, broccoli, cauliflower, kale, collard greens, and Brussels sprouts: they are all bred from, and are the same species as, the wild cabbage plant (Brassica oleracea). The taste and texture of kohlrabi are similar to those of a broccoli stem or cabbage heart, but milder and sweeter, with a higher ratio of flesh to skin. The young stem in particular can be as crisp and juicy as an apple, although much less sweet. Except for the Gigante cultivar, spring-grown kohlrabi much over 5 cm in size tend to be woody, as do full-grown kohlrabi much over perhaps 10 cm in size; the Gigante cultivar can achieve great size while remaining of good eating quality. The plant matures in 55–60 days after sowing and has good standing ability for up to 30 days after maturity. The approximate weight is 150 g. It grows well in hydroponic systems, producing a large edible bulk without clogging the nutrient troughs. There are several varieties commonly available, including 'White Vienna', 'Purple Vienna', 'Grand Duke', 'Gigante' (also known as "Superschmelz"), 'Purple Danube', and 'White Danube'. Coloration of the purple types is superficial: the edible parts are all pale yellow. The leafy greens can also be eaten. One commonly used variety grows without a swollen stem, having just leaves and a very thin stem, and is called Haakh. Haakh and Monj are popular Kashmiri dishes made using this vegetable. In the second year, the plant will bloom and develop seeds. Preparation and use Kohlrabi stems (the enlarged vegetal part) are surrounded by two distinct fibrous layers that do not soften appreciably when cooked. These layers are generally peeled away prior to cooking or serving raw, with the result that the stems often provide a smaller amount of food than one might assume from their intact appearance. The bulbous kohlrabi stem is frequently used raw in salad or slaws. It has a texture similar to that of a broccoli stem, but with a flavor that is sweeter and less vegetal. Kohlrabi leaves are edible and can be used similarly to collard greens and kale, but take longer to cook. Kohlrabi is an important part of Kashmiri cuisine, where it is called Mŏnji. It is one of the most commonly cooked vegetables, along with collard greens (haakh). It is prepared with its leaves and served with a light soup and eaten with rice. In Cyprus, it is popularly sprinkled with salt and lemon and served as an appetizer. Kohlrabi is a common ingredient in Vietnamese cuisine. It can also be found in the dish nem rán, stir fry and canh. Raw kohlrabi is usually sliced thinly for nộm or nước chấm. Some varieties are grown as feed for cattle. References What is Kohlrabi? External links PROTAbase on Brassica oleracea (kohlrabi) Horticultural information on the tasty kohlrabi From the Learn2Grow databases Kohlrabi and Brussels Sprouts Are European Brassica oleracea Stem vegetables